Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
604,401
18,683,861,383
IssuesEvent
2021-11-01 09:52:39
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
opened
[UI Mobile] The club's cover photo is not aligned on 'Club's card' page
bug UI Priority: Low
**Environment**: Samsung Edge S7 **Reproducible**: always **Build found**: the last commit **Steps to reproduce** 1. Log in to https://speak-ukrainian.org.ua/dev with valid data. 2. Click on the 'Menu' button for website navigation and choose 'Гуртки' tab 3. Choose any club showing on the 'Club list' page and click 'Детальніше'. 4. Verify the alignment of the club's cover photo does not correspond to the mock-up **Actual result** The alignment of the club's cover photo on 'Club's card' page does not correspond to the mock-up <img width="184" alt="Screenshot_28" src="https://user-images.githubusercontent.com/91254317/139653794-d08986b2-1dc5-4904-bc6e-3842ca8ca5dc.png"> **Expected result** The alignment of the club's cover photo on 'Club's card' page corresponds to the mock-up <img width="225" alt="gggggggggggggggggggggg" src="https://user-images.githubusercontent.com/91254317/139653776-63bba40d-792b-4f8b-a751-d6c0504d9ca9.png">
1.0
[UI Mobile] The club's cover photo is not aligned on 'Club's card' page - **Environment**: Samsung Edge S7 **Reproducible**: always **Build found**: the last commit **Steps to reproduce** 1. Log in to https://speak-ukrainian.org.ua/dev with valid data. 2. Click on the 'Menu' button for website navigation and choose 'Гуртки' tab 3. Choose any club showing on the 'Club list' page and click 'Детальніше'. 4. Verify the alignment of the club's cover photo does not correspond to the mock-up **Actual result** The alignment of the club's cover photo on 'Club's card' page does not correspond to the mock-up <img width="184" alt="Screenshot_28" src="https://user-images.githubusercontent.com/91254317/139653794-d08986b2-1dc5-4904-bc6e-3842ca8ca5dc.png"> **Expected result** The alignment of the club's cover photo on 'Club's card' page corresponds to the mock-up <img width="225" alt="gggggggggggggggggggggg" src="https://user-images.githubusercontent.com/91254317/139653776-63bba40d-792b-4f8b-a751-d6c0504d9ca9.png">
priority
the club s cover photo is not aligned on club s card page environment samsung edge reproducible always build found the last commit steps to reproduce log in to with valid data click on the menu button for website navigation and choose гуртки tab choose any club showing on the club list page and click детальніше verify the alignment of the club s cover photo does not correspond to the mock up actual result the alignment of the club s cover photo on club s card page does not correspond to the mock up img width alt screenshot src expected result the alignment of the club s cover photo on club s card page corresponds to the mock up img width alt gggggggggggggggggggggg src
1
42,514
22,638,869,589
IssuesEvent
2022-06-30 22:18:59
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
[Impeller] New gallery transition perf worst and 99th %-ile raster time regression on Android
engine severe: performance perf: speed P4 impeller
On the change to render BackdropFilter layers directly to DisplayListBuilder https://github.com/flutter/engine/pull/34337, worst and 99th %-ile frame rasterization times regressed on Moto G4: https://flutter-flutter-perf.skia.org/e/?begin=1654156387&end=1656469887&keys=X5f6e343cfb7467cdc34ede6764f63458&requestType=0&xbaroffset=29603 @chinmaygarde @bdero @flar
True
[Impeller] New gallery transition perf worst and 99th %-ile raster time regression on Android - On the change to render BackdropFilter layers directly to DisplayListBuilder https://github.com/flutter/engine/pull/34337, worst and 99th %-ile frame rasterization times regressed on Moto G4: https://flutter-flutter-perf.skia.org/e/?begin=1654156387&end=1656469887&keys=X5f6e343cfb7467cdc34ede6764f63458&requestType=0&xbaroffset=29603 @chinmaygarde @bdero @flar
non_priority
new gallery transition perf worst and ile raster time regression on android on the change to render backdropfilter layers directly to displaylistbuilder worst and ile frame rasterization times regressed on moto chinmaygarde bdero flar
0
406,684
27,577,103,333
IssuesEvent
2023-03-08 13:49:40
aesim-tech/simba-project
https://api.github.com/repos/aesim-tech/simba-project
opened
Fix Thermal Data documentation
documentation
**Is your feature request related to a problem? Please describe.** ![image](https://user-images.githubusercontent.com/112709259/223729124-06d186aa-380b-464e-87b4-fbe1e1815083.png) There is an error in this picture on the Foster chain. It is impossible to connect a Foster network to a Cauer network (heat sink) as shown in the figure. It is necessary to connect a cauer network with another cauer or a Foster network with a Foster network. But do not mix the two as shown here. **Describe the solution you'd like** We should fix the image.
1.0
Fix Thermal Data documentation - **Is your feature request related to a problem? Please describe.** ![image](https://user-images.githubusercontent.com/112709259/223729124-06d186aa-380b-464e-87b4-fbe1e1815083.png) There is an error in this picture on the Foster chain. It is impossible to connect a Foster network to a Cauer network (heat sink) as shown in the figure. It is necessary to connect a cauer network with another cauer or a Foster network with a Foster network. But do not mix the two as shown here. **Describe the solution you'd like** We should fix the image.
non_priority
fix thermal data documentation is your feature request related to a problem please describe there is an error in this picture on the foster chain it is impossible to connect a foster network to a cauer network heat sink as shown in the figure it is necessary to connect a cauer network with another cauer or a foster network with a foster network but do not mix the two as shown here describe the solution you d like we should fix the image
0
79,914
9,966,327,335
IssuesEvent
2019-07-08 10:53:44
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
vscode-uri escapes colon in drive letter
*as-designed under-discussion uri
We are trying to use vscode-uri in the debug adapter, and one issue is that vscode-uri uri-encodes the colon after the drive letter in a windows file uri which causes issues downstream. I see https://github.com/microsoft/vscode/issues/2990#issuecomment-204295374 but I'm worried that will just cause other issues by not encoding other characters. And I'm wondering why we do this by default, this file URI scheme spec calls out `file:///c:/foo` as being a valid file URI. https://tools.ietf.org/html/rfc8089
1.0
vscode-uri escapes colon in drive letter - We are trying to use vscode-uri in the debug adapter, and one issue is that vscode-uri uri-encodes the colon after the drive letter in a windows file uri which causes issues downstream. I see https://github.com/microsoft/vscode/issues/2990#issuecomment-204295374 but I'm worried that will just cause other issues by not encoding other characters. And I'm wondering why we do this by default, this file URI scheme spec calls out `file:///c:/foo` as being a valid file URI. https://tools.ietf.org/html/rfc8089
non_priority
vscode uri escapes colon in drive letter we are trying to use vscode uri in the debug adapter and one issue is that vscode uri uri encodes the colon after the drive letter in a windows file uri which causes issues downstream i see but i m worried that will just cause other issues by not encoding other characters and i m wondering why we do this by default this file uri scheme spec calls out file c foo as being a valid file uri
0
95,648
3,954,704,922
IssuesEvent
2016-04-29 18:01:35
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
opened
Ingesting PortAuPrinceOsmPoiRoadBuilding.osm file results in cache error when exportrenderdb script is run
Category: Core Priority: Medium Status: Defined Type: Bug
2016-04-29 13:59:21,370 ERROR JobExecutionManager:225 - Job with ID: 70c4bae4-666e-45e4-8cfd-595428e7dd8a failed: Failed to execute.Error running osm2ogr: Relation element with ID: 11380 and type: Way did not exist in the element cache with size = 20000. make: *** [step1] Error 255 2016-04-29 13:59:21,380 ERROR RasterToTilesService:48 - Failure ingesting resource hoot.services.nativeInterfaces.NativeInterfaceException: Failed to execute.Error running osm2ogr: Relation element with ID: 11380 and type: Way did not exist in the element cache with size = 20000. make: *** [step1] Error 255
1.0
Ingesting PortAuPrinceOsmPoiRoadBuilding.osm file results in cache error when exportrenderdb script is run - 2016-04-29 13:59:21,370 ERROR JobExecutionManager:225 - Job with ID: 70c4bae4-666e-45e4-8cfd-595428e7dd8a failed: Failed to execute.Error running osm2ogr: Relation element with ID: 11380 and type: Way did not exist in the element cache with size = 20000. make: *** [step1] Error 255 2016-04-29 13:59:21,380 ERROR RasterToTilesService:48 - Failure ingesting resource hoot.services.nativeInterfaces.NativeInterfaceException: Failed to execute.Error running osm2ogr: Relation element with ID: 11380 and type: Way did not exist in the element cache with size = 20000. make: *** [step1] Error 255
priority
ingesting portauprinceosmpoiroadbuilding osm file results in cache error when exportrenderdb script is run error jobexecutionmanager job with id failed failed to execute error running relation element with id and type way did not exist in the element cache with size make error error rastertotilesservice failure ingesting resource hoot services nativeinterfaces nativeinterfaceexception failed to execute error running relation element with id and type way did not exist in the element cache with size make error
1
121,651
17,661,987,133
IssuesEvent
2021-08-21 17:50:48
ghc-dev/Dale-Park
https://api.github.com/repos/ghc-dev/Dale-Park
opened
CVE-2021-3533 (Low) detected in ansible-2.9.9.tar.gz
security vulnerability
## CVE-2021-3533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Dale-Park/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Dale-Park/commit/c16e5eddfe2d758a65db13af89bc7f384b372ec2">c16e5eddfe2d758a65db13af89bc7f384b372ec2</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR=/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533>CVE-2021-3533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3533","vulnerabilityDetails":"Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR\u003d/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data.\n","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533","cvss3Severity":"low","cvss3Score":"2.5","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-3533 (Low) detected in ansible-2.9.9.tar.gz - ## CVE-2021-3533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Dale-Park/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Dale-Park/commit/c16e5eddfe2d758a65db13af89bc7f384b372ec2">c16e5eddfe2d758a65db13af89bc7f384b372ec2</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR=/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533>CVE-2021-3533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3533","vulnerabilityDetails":"Vulnerability in ansible when ANSIBLE_ASYNC_DIR defaults to ~/.ansible_async/ but is settable by the user. It can be set by the ansible user to a subdirectory of a world writable directory, for instance ANSIBLE_ASYNC_DIR\u003d/tmp/username-ansible-async/. When this occurs, there is a race condition on the managed machine. A malicious, low privileged account on the remote machine can pre-create /tmp/username-ansible-async and then use various attacks to access the async result data.\n","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3533","cvss3Severity":"low","cvss3Score":"2.5","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve low detected in ansible tar gz cve low severity vulnerability vulnerable library ansible tar gz radically simple it automation library home page a href path to dependency file dale park requirements txt path to vulnerable library requirements txt dependency hierarchy x ansible tar gz vulnerable library found in head commit a href found in base branch master vulnerability details vulnerability in ansible when ansible async dir defaults to ansible async but is settable by the user it can be set by the ansible user to a subdirectory of a world writable directory for instance ansible async dir tmp username ansible async when this occurs there is a race condition on the managed machine a malicious low privileged account on the remote machine can pre create tmp username ansible async and then use various attacks to access the async result data publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree ansible isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in ansible when ansible async dir defaults to ansible async but is settable by the user it can be set by the ansible user to a subdirectory of a world writable directory for instance ansible async dir tmp username ansible async when this occurs there is a race condition on the managed machine a malicious low privileged account on the remote machine can pre create tmp username ansible async and then use various attacks to access the async result data n vulnerabilityurl
0
326,421
24,083,730,635
IssuesEvent
2022-09-19 09:02:59
valory-xyz/open-aea
https://api.github.com/repos/valory-xyz/open-aea
closed
Docs Generator No longer generates ABC class docs.
bug documentation
The script is; ./scripts/generate_api_docs The relevant libraries are; pydoc-markdown https://niklasrosenstein.github.io/pydoc-markdown/ Which uses; https://niklasrosenstein.github.io/docspec/ I have been unable to identify anything within both of these packages which would allow us to generate the ABCs
1.0
Docs Generator No longer generates ABC class docs. - The script is; ./scripts/generate_api_docs The relevant libraries are; pydoc-markdown https://niklasrosenstein.github.io/pydoc-markdown/ Which uses; https://niklasrosenstein.github.io/docspec/ I have been unable to identify anything within both of these packages which would allow us to generate the ABCs
non_priority
docs generator no longer generates abc class docs the script is scripts generate api docs the relevant libraries are pydoc markdown which uses i have been unable to identify anything within both of these packages which would allow us to generate the abcs
0
227,266
18,054,249,846
IssuesEvent
2021-09-20 05:20:31
logicmoo/logicmoo_workspace
https://api.github.com/repos/logicmoo/logicmoo_workspace
opened
logicmoo.pfc.test.sanity_base.VERBATUMS_04 JUnit
Test_9999 logicmoo.pfc.test.sanity_base unit_test VERBATUMS_04 Failing
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif verbatums_04.pfc) % ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/ % EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/verbatums_04.pfc % JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/VERBATUMS_04/logicmoo_pfc_test_sanity_base_VERBATUMS_04_JUnit/ % ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AVERBATUMS_04 ``` %~ init_phase(after_load) %~ init_phase(restore_state) % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/verbatums_04.pfc'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- dynamic(total_income/3). % RULES % RULES income(Person,_Source,Year,Dollars) ==> {increment_income(Person,Year,Dollars)}. ==> do_and_undo(increment_income(P,Y,D),decrement_income(P,Y,D)). increment_income(P,Y,D) :- (retract(total_income(P,Y,Old)) -> New is Old+D ; New = D), assert(total_income(P,Y,New)). decrement_income(P,Y,D) :- retract(total_income(P,Y,Old)), New is Old-D, assert(total_income(P,Y,New)). % FACTS % FACTS income(person,sourceOne,2035,6666). income(person,sourceTwo,2035,1111). income(person,sourceTwo,2036,2222). % RESULTS PT 1 % RESULTS PT 1 :- listing(total_income/3). %~ skipped( listing( total_income/3)) /* total_income(person, 2035, 7777). total_income(person, 2036, 2222). */ % UPDATE Remove some income \+ income(person,_,2035,1111). % RESULTS PT 2 %~ debugm( baseKB, %~ show_success( baseKB, %~ baseKB : mpred_withdraw( income(person,sourceTwo,2035,1111), %~ ( mfl4(BaseKB,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/edit/master/packs_sys/pfc/t/sanity_base/verbatums_04.pfc ',50) , %~ ax)))) % RESULTS PT 2 :- listing(total_income/3). %~ skipped( listing( total_income/3)) %~ unused(no_junit_results) %~ test_completed_exit(0) ``` totalTime=1.000 FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k verbatums_04.pfc (returned 0) Add_LABELS='' Rem_LABELS='Skipped,Skipped,Errors,Warnings,Overtime,Skipped,Skipped'
3.0
logicmoo.pfc.test.sanity_base.VERBATUMS_04 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s swipl -x /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-clif verbatums_04.pfc) % ISSUE: https://github.com/logicmoo/logicmoo_workspace/issues/ % EDIT: https://github.com/logicmoo/logicmoo_workspace/edit/master/packs_sys/pfc/t/sanity_base/verbatums_04.pfc % JENKINS: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/VERBATUMS_04/logicmoo_pfc_test_sanity_base_VERBATUMS_04_JUnit/ % ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3AVERBATUMS_04 ``` %~ init_phase(after_load) %~ init_phase(restore_state) % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/verbatums_04.pfc'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- dynamic(total_income/3). % RULES % RULES income(Person,_Source,Year,Dollars) ==> {increment_income(Person,Year,Dollars)}. ==> do_and_undo(increment_income(P,Y,D),decrement_income(P,Y,D)). increment_income(P,Y,D) :- (retract(total_income(P,Y,Old)) -> New is Old+D ; New = D), assert(total_income(P,Y,New)). decrement_income(P,Y,D) :- retract(total_income(P,Y,Old)), New is Old-D, assert(total_income(P,Y,New)). % FACTS % FACTS income(person,sourceOne,2035,6666). income(person,sourceTwo,2035,1111). income(person,sourceTwo,2036,2222). % RESULTS PT 1 % RESULTS PT 1 :- listing(total_income/3). %~ skipped( listing( total_income/3)) /* total_income(person, 2035, 7777). total_income(person, 2036, 2222). */ % UPDATE Remove some income \+ income(person,_,2035,1111). % RESULTS PT 2 %~ debugm( baseKB, %~ show_success( baseKB, %~ baseKB : mpred_withdraw( income(person,sourceTwo,2035,1111), %~ ( mfl4(BaseKB,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/edit/master/packs_sys/pfc/t/sanity_base/verbatums_04.pfc ',50) , %~ ax)))) % RESULTS PT 2 :- listing(total_income/3). %~ skipped( listing( total_income/3)) %~ unused(no_junit_results) %~ test_completed_exit(0) ``` totalTime=1.000 FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k verbatums_04.pfc (returned 0) Add_LABELS='' Rem_LABELS='Skipped,Skipped,Errors,Warnings,Overtime,Skipped,Skipped'
non_priority
logicmoo pfc test sanity base verbatums junit cd var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base timeout foreground preserve status s sigkill k swipl x var lib jenkins workspace logicmoo workspace bin lmoo clif verbatums pfc issue edit jenkins issue search init phase after load init phase restore state running var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base verbatums pfc this test might need use module library logicmoo plarkc dynamic total income rules rules income person source year dollars increment income person year dollars do and undo increment income p y d decrement income p y d increment income p y d retract total income p y old new is old d new d assert total income p y new decrement income p y d retract total income p y old new is old d assert total income p y new facts facts income person sourceone income person sourcetwo income person sourcetwo results pt results pt listing total income skipped listing total income total income person total income person update remove some income income person results pt debugm basekb show success basekb basekb mpred withdraw income person sourcetwo basekb basekb ax results pt listing total income skipped listing total income unused no junit results test completed exit totaltime failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k verbatums pfc returned add labels rem labels skipped skipped errors warnings overtime skipped skipped
0
34,783
14,520,870,808
IssuesEvent
2020-12-14 06:21:23
MicrosoftDocs/dynamics-365-customer-engagement
https://api.github.com/repos/MicrosoftDocs/dynamics-365-customer-engagement
closed
Sample?
Channel Integration Framework Pri2 assigned-to-author dynamics-365-customerservice/svc waiting on customer
Any sample code on how to raise the onclicktoact event? I've tried every way I know how --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 829606f3-febb-1082-5233-fc754fe79bbe * Version Independent ID: be5e2a6d-674d-c598-c667-6445917a4409 * Content: [onclicktoact event (JavaScript API Reference) for Dynamics 365 Channel Integration Framework (CIF) version 1.0](https://docs.microsoft.com/en-us/dynamics365/customer-service/channel-integration-framework/reference/events/onclicktoact) * Content Source: [ce/customer-service/channel-integration-framework/reference/events/onclicktoact.md](https://github.com/MicrosoftDocs/dynamics-365-customer-engagement/blob/master/ce/customer-service/channel-integration-framework/reference/events/onclicktoact.md) * Service: **dynamics-365-customerservice** * GitHub Login: @susikka * Microsoft Alias: **susikka**
1.0
Sample? - Any sample code on how to raise the onclicktoact event? I've tried every way I know how --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 829606f3-febb-1082-5233-fc754fe79bbe * Version Independent ID: be5e2a6d-674d-c598-c667-6445917a4409 * Content: [onclicktoact event (JavaScript API Reference) for Dynamics 365 Channel Integration Framework (CIF) version 1.0](https://docs.microsoft.com/en-us/dynamics365/customer-service/channel-integration-framework/reference/events/onclicktoact) * Content Source: [ce/customer-service/channel-integration-framework/reference/events/onclicktoact.md](https://github.com/MicrosoftDocs/dynamics-365-customer-engagement/blob/master/ce/customer-service/channel-integration-framework/reference/events/onclicktoact.md) * Service: **dynamics-365-customerservice** * GitHub Login: @susikka * Microsoft Alias: **susikka**
non_priority
sample any sample code on how to raise the onclicktoact event i ve tried every way i know how document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id febb version independent id content content source service dynamics customerservice github login susikka microsoft alias susikka
0
619,135
19,517,167,784
IssuesEvent
2021-12-29 12:16:19
slcommunity/til-back
https://api.github.com/repos/slcommunity/til-back
closed
FormatDatetime 설정
Feature/Function Status:To Do Priority:High
## 목적 작성일에서 Date를 다 찍지말고 format 정하기 ## 작업 상세 내용 - [x] (YYYY-MM-DD) ## 참고사항
1.0
FormatDatetime 설정 - ## 목적 작성일에서 Date를 다 찍지말고 format 정하기 ## 작업 상세 내용 - [x] (YYYY-MM-DD) ## 참고사항
priority
formatdatetime 설정 목적 작성일에서 date를 다 찍지말고 format 정하기 작업 상세 내용 yyyy mm dd 참고사항
1
40,624
5,243,298,867
IssuesEvent
2017-01-31 20:20:00
verfriemelt-dot-org/sachsencacher
https://api.github.com/repos/verfriemelt-dot-org/sachsencacher
closed
listensuche ungünstig plaziert.
Design enhancement
wir müssen mal die suchleiste besser positionieren auf den listen seiten. die gefilterte suche steht in konkurrenz mit der globalen suche und verwirrt an der stelle etwas, das muss vom design her klar sein, um was es da geht.
1.0
listensuche ungünstig plaziert. - wir müssen mal die suchleiste besser positionieren auf den listen seiten. die gefilterte suche steht in konkurrenz mit der globalen suche und verwirrt an der stelle etwas, das muss vom design her klar sein, um was es da geht.
non_priority
listensuche ungünstig plaziert wir müssen mal die suchleiste besser positionieren auf den listen seiten die gefilterte suche steht in konkurrenz mit der globalen suche und verwirrt an der stelle etwas das muss vom design her klar sein um was es da geht
0
380,140
11,254,273,024
IssuesEvent
2020-01-11 22:18:50
python-discord/bot
https://api.github.com/repos/python-discord/bot
opened
PEP Command Fails on Empty Metadata
area: information priority: 3 - low type: bug
Currently, the PEP command assumes that if a metadata field is present in the PEP's summary table that the value is populated: https://github.com/python-discord/bot/blob/74d990540a1072c1782fa7593d7d1abe3c165f49/bot/cogs/utils.py#L64-L72 However, this is not always the case, e.g. [PEP 249](https://www.python.org/dev/peps/pep-0249/): ![image](https://user-images.githubusercontent.com/5323929/72211253-23e26600-3496-11ea-847b-29300866532b.png) Because of this, the embed field `value` is provided a null value, causing an error to be raised: ``` HTTPException: 400 BAD REQUEST (error code: 50035): Invalid Form Body In embed.fields.1.value: This field is required ``` To fix this, there should be a catch for empty values to prevent a field being added if there's no value to populate.
1.0
PEP Command Fails on Empty Metadata - Currently, the PEP command assumes that if a metadata field is present in the PEP's summary table that the value is populated: https://github.com/python-discord/bot/blob/74d990540a1072c1782fa7593d7d1abe3c165f49/bot/cogs/utils.py#L64-L72 However, this is not always the case, e.g. [PEP 249](https://www.python.org/dev/peps/pep-0249/): ![image](https://user-images.githubusercontent.com/5323929/72211253-23e26600-3496-11ea-847b-29300866532b.png) Because of this, the embed field `value` is provided a null value, causing an error to be raised: ``` HTTPException: 400 BAD REQUEST (error code: 50035): Invalid Form Body In embed.fields.1.value: This field is required ``` To fix this, there should be a catch for empty values to prevent a field being added if there's no value to populate.
priority
pep command fails on empty metadata currently the pep command assumes that if a metadata field is present in the pep s summary table that the value is populated however this is not always the case e g because of this the embed field value is provided a null value causing an error to be raised httpexception bad request error code invalid form body in embed fields value this field is required to fix this there should be a catch for empty values to prevent a field being added if there s no value to populate
1
730,201
25,163,936,830
IssuesEvent
2022-11-10 19:03:59
zowe/vscode-extension-for-zowe
https://api.github.com/repos/zowe/vscode-extension-for-zowe
closed
ZE: create all `.zowe` subfolders needed (e.g. `~/.zowe/settings`)
priority-high 22PI4
We would still want to have the means of creating all other missing folders (`~/.zowe/settings`) _Originally posted by @zFernand0 in https://github.com/zowe/vscode-extension-for-zowe/issues/1830#issuecomment-1161834205_
1.0
ZE: create all `.zowe` subfolders needed (e.g. `~/.zowe/settings`) - We would still want to have the means of creating all other missing folders (`~/.zowe/settings`) _Originally posted by @zFernand0 in https://github.com/zowe/vscode-extension-for-zowe/issues/1830#issuecomment-1161834205_
priority
ze create all zowe subfolders needed e g zowe settings we would still want to have the means of creating all other missing folders zowe settings originally posted by in
1
797,491
28,147,217,394
IssuesEvent
2023-04-02 16:17:30
x13pixels/remedybg-issues
https://api.github.com/repos/x13pixels/remedybg-issues
closed
`char8_t` support
Type: Bug Priority: 6 (Medium) Component: Watch Window
Currently, RemedyBG displays a `char8_t` array as a `TBI []` type and this makes it hard to debug in certain cases (it doesn't seem to recognize `static` allocation of these in the global variables view).
1.0
`char8_t` support - Currently, RemedyBG displays a `char8_t` array as a `TBI []` type and this makes it hard to debug in certain cases (it doesn't seem to recognize `static` allocation of these in the global variables view).
priority
t support currently remedybg displays a t array as a tbi type and this makes it hard to debug in certain cases it doesn t seem to recognize static allocation of these in the global variables view
1
95,646
10,884,649,787
IssuesEvent
2019-11-18 08:47:47
emielvanseveren/hyperledger
https://api.github.com/repos/emielvanseveren/hyperledger
closed
Word template (project mgmt)
documentation
Het aanmaken van een > mooi template waar we de project analyse in kunnen steken.
1.0
Word template (project mgmt) - Het aanmaken van een > mooi template waar we de project analyse in kunnen steken.
non_priority
word template project mgmt het aanmaken van een mooi template waar we de project analyse in kunnen steken
0
93,981
11,839,612,492
IssuesEvent
2020-03-23 17:26:03
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Predictive typing incorrectly performs selection when dot character is typed
*as-designed
Issue Type: <b>Bug</b> When I start typing, "vue" the following predictive typing popup is display. ![bug](https://user-images.githubusercontent.com/439226/77030156-2e2c4f00-6974-11ea-9b43-811258e3e29f.png) The problem is when I type ".", vscode incorrectly replaces the text "vue" with "VTTCue." which is not what I expect. What I expect is, "__vue.__" VS Code version: Code - Insiders 1.43.0-insider (e6a45f4242ebddb7aa9a229f85555e8a3bd987e2, 2020-02-21T05:40:52.685Z) OS version: Linux x64 5.5.7-1-default <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|AMD Ryzen 7 2700X Eight-Core Processor (16 x 4051)| |GPU Status|2d_canvas: enabled<br>flash_3d: enabled<br>flash_stage3d: enabled<br>flash_stage3d_baseline: enabled<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>oop_rasterization: disabled_off<br>protected_video_decode: unavailable_off<br>rasterization: disabled_software<br>skia_renderer: disabled_off_ok<br>video_decode: unavailable_off<br>viz_display_compositor: enabled_on<br>viz_hit_test_surface_layer: disabled_off_ok<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|0, 1, 1| |Memory (System)|31.30GB (1.35GB free)| |Process Argv|. --no-sandbox| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (26)</summary> Extension|Author (truncated)|Version ---|---|--- terraform|4op|0.1.8 vscode-base64|ada|0.1.0 Bookmarks|ale|11.0.0 ng-template|Ang|0.900.18 npm-intellisense|chr|1.3.0 vscode-markdownlint|Dav|0.34.0 vscode-eslint|dba|2.1.1 EditorConfig|Edi|0.14.4 prettier-vscode|esb|3.20.0 shell-format|fox|7.0.1 markdown-preview-bitbucket|hbr|1.0.0 vscode-peacock|joh|3.6.0 solidity|Jua|0.0.70 sort-typescript-imports|mic|1.4.1 vscode-docker|ms-|1.0.0 python|ms-|2020.3.64983-dev cpptools|ms-|0.27.0-insiders2 vscode-typescript-tslint-plugin|ms-|1.2.3 debugger-for-chrome|msj|4.12.6 angular2-inline|nat|0.0.17 vscode-yaml|red|0.7.2 markdown-preview-enhanced|shd|0.5.2 language-stylus|sys|1.11.0 sort-lines|Tyr|1.9.0 better-align|wwm|1.1.6 markdown-pdf|yza|1.4.4 (5 theme extensions excluded) </details> <!-- generated by issue reporter -->
1.0
Predictive typing incorrectly performs selection when dot character is typed - Issue Type: <b>Bug</b> When I start typing, "vue" the following predictive typing popup is display. ![bug](https://user-images.githubusercontent.com/439226/77030156-2e2c4f00-6974-11ea-9b43-811258e3e29f.png) The problem is when I type ".", vscode incorrectly replaces the text "vue" with "VTTCue." which is not what I expect. What I expect is, "__vue.__" VS Code version: Code - Insiders 1.43.0-insider (e6a45f4242ebddb7aa9a229f85555e8a3bd987e2, 2020-02-21T05:40:52.685Z) OS version: Linux x64 5.5.7-1-default <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|AMD Ryzen 7 2700X Eight-Core Processor (16 x 4051)| |GPU Status|2d_canvas: enabled<br>flash_3d: enabled<br>flash_stage3d: enabled<br>flash_stage3d_baseline: enabled<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>oop_rasterization: disabled_off<br>protected_video_decode: unavailable_off<br>rasterization: disabled_software<br>skia_renderer: disabled_off_ok<br>video_decode: unavailable_off<br>viz_display_compositor: enabled_on<br>viz_hit_test_surface_layer: disabled_off_ok<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|0, 1, 1| |Memory (System)|31.30GB (1.35GB free)| |Process Argv|. --no-sandbox| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (26)</summary> Extension|Author (truncated)|Version ---|---|--- terraform|4op|0.1.8 vscode-base64|ada|0.1.0 Bookmarks|ale|11.0.0 ng-template|Ang|0.900.18 npm-intellisense|chr|1.3.0 vscode-markdownlint|Dav|0.34.0 vscode-eslint|dba|2.1.1 EditorConfig|Edi|0.14.4 prettier-vscode|esb|3.20.0 shell-format|fox|7.0.1 markdown-preview-bitbucket|hbr|1.0.0 vscode-peacock|joh|3.6.0 solidity|Jua|0.0.70 sort-typescript-imports|mic|1.4.1 vscode-docker|ms-|1.0.0 python|ms-|2020.3.64983-dev cpptools|ms-|0.27.0-insiders2 vscode-typescript-tslint-plugin|ms-|1.2.3 debugger-for-chrome|msj|4.12.6 angular2-inline|nat|0.0.17 vscode-yaml|red|0.7.2 markdown-preview-enhanced|shd|0.5.2 language-stylus|sys|1.11.0 sort-lines|Tyr|1.9.0 better-align|wwm|1.1.6 markdown-pdf|yza|1.4.4 (5 theme extensions excluded) </details> <!-- generated by issue reporter -->
non_priority
predictive typing incorrectly performs selection when dot character is typed issue type bug when i start typing vue the following predictive typing popup is display the problem is when i type vscode incorrectly replaces the text vue with vttcue which is not what i expect what i expect is vue vs code version code insiders insider os version linux default system info item value cpus amd ryzen eight core processor x gpu status canvas enabled flash enabled flash enabled flash baseline enabled gpu compositing enabled multiple raster threads enabled on oop rasterization disabled off protected video decode unavailable off rasterization disabled software skia renderer disabled off ok video decode unavailable off viz display compositor enabled on viz hit test surface layer disabled off ok webgl enabled enabled load avg memory system free process argv no sandbox screen reader no vm extensions extension author truncated version terraform vscode ada bookmarks ale ng template ang npm intellisense chr vscode markdownlint dav vscode eslint dba editorconfig edi prettier vscode esb shell format fox markdown preview bitbucket hbr vscode peacock joh solidity jua sort typescript imports mic vscode docker ms python ms dev cpptools ms vscode typescript tslint plugin ms debugger for chrome msj inline nat vscode yaml red markdown preview enhanced shd language stylus sys sort lines tyr better align wwm markdown pdf yza theme extensions excluded
0
477,427
13,762,153,615
IssuesEvent
2020-10-07 08:45:14
HackYourFuture-CPH/chattie
https://api.github.com/repos/HackYourFuture-CPH/chattie
closed
When signing up, add the username to the database
High priority User story
There is a bug now, where the username is not added to the database
1.0
When signing up, add the username to the database - There is a bug now, where the username is not added to the database
priority
when signing up add the username to the database there is a bug now where the username is not added to the database
1
48,086
2,990,153,033
IssuesEvent
2015-07-21 07:21:20
jayway/rest-assured
https://api.github.com/repos/jayway/rest-assured
closed
Replace host:port to baseUri in log when path contains query
bug imported Priority-Medium
_From [twilig...@gmail.com](https://code.google.com/u/115210657126813255036/) on January 30, 2014 11:01:31_ What steps will reproduce the problem? given().log().all().expect().log().all().get(" http://ya.ru/bla/?param"); What is the expected output? What do you see instead? Expect in log: Request method: GET Request path: http://ya.ru/bla/?param Request params: \<none> Query params: \<none> Form params: \<none> Path params: \<none> Headers: Content-Type=*/* Cookies: \<none> Body: \<none> Got: Request method: GET Request path: http://localhost:8080/bla/?param Request params: \<none> Query params: \<none> Form params: \<none> Path params: \<none> Headers: Content-Type=*/* Cookies: \<none> Body: \<none> What version of the product are you using? On what operating system? 2.3.0, any OS Please provide any additional information below. Error in: com.jayway.restassured.internal.RequestSpecificationImpl#generateRequestUriToLog When we don't have query, we do (path is : http://ya.ru/bla/?param , targetPath is: only /bla/): } else { targetPath = path } then use it to build log uri. And when we have query, we send to build uri only path without host and after that host is baseUri host. https://groups.google.com/forum/#!topic/rest-assured/IWhu34xL3rQ _Original issue: http://code.google.com/p/rest-assured/issues/detail?id=304_
1.0
Replace host:port to baseUri in log when path contains query - _From [twilig...@gmail.com](https://code.google.com/u/115210657126813255036/) on January 30, 2014 11:01:31_ What steps will reproduce the problem? given().log().all().expect().log().all().get(" http://ya.ru/bla/?param"); What is the expected output? What do you see instead? Expect in log: Request method: GET Request path: http://ya.ru/bla/?param Request params: \<none> Query params: \<none> Form params: \<none> Path params: \<none> Headers: Content-Type=*/* Cookies: \<none> Body: \<none> Got: Request method: GET Request path: http://localhost:8080/bla/?param Request params: \<none> Query params: \<none> Form params: \<none> Path params: \<none> Headers: Content-Type=*/* Cookies: \<none> Body: \<none> What version of the product are you using? On what operating system? 2.3.0, any OS Please provide any additional information below. Error in: com.jayway.restassured.internal.RequestSpecificationImpl#generateRequestUriToLog When we don't have query, we do (path is : http://ya.ru/bla/?param , targetPath is: only /bla/): } else { targetPath = path } then use it to build log uri. And when we have query, we send to build uri only path without host and after that host is baseUri host. https://groups.google.com/forum/#!topic/rest-assured/IWhu34xL3rQ _Original issue: http://code.google.com/p/rest-assured/issues/detail?id=304_
priority
replace host port to baseuri in log when path contains query from on january what steps will reproduce the problem given log all expect log all get what is the expected output what do you see instead expect in log request method get request path request params query params form params path params headers content type cookies body got request method get request path request params query params form params path params headers content type cookies body what version of the product are you using on what operating system any os please provide any additional information below error in com jayway restassured internal requestspecificationimpl generaterequesturitolog when we don t have query we do path is targetpath is only bla else targetpath path then use it to build log uri and when we have query we send to build uri only path without host and after that host is baseuri host original issue
1
768,271
26,959,907,444
IssuesEvent
2023-02-08 17:22:20
Automattic/woocommerce-payments
https://api.github.com/repos/Automattic/woocommerce-payments
opened
Split UPE: migrate saved SEPA tokens
type: bug priority: high
### Describe the bug <!-- A clear and concise description of what the bug is. Please be as descriptive as possible. --> If a merchant switches from the "Legacy" UPE to the "Split UPE", saved SEPA tokens will not appear anymore. We need to find a way to migrate the SEPA tokens to the new SEPA gateway.
1.0
Split UPE: migrate saved SEPA tokens - ### Describe the bug <!-- A clear and concise description of what the bug is. Please be as descriptive as possible. --> If a merchant switches from the "Legacy" UPE to the "Split UPE", saved SEPA tokens will not appear anymore. We need to find a way to migrate the SEPA tokens to the new SEPA gateway.
priority
split upe migrate saved sepa tokens describe the bug if a merchant switches from the legacy upe to the split upe saved sepa tokens will not appear anymore we need to find a way to migrate the sepa tokens to the new sepa gateway
1
368,122
10,866,426,066
IssuesEvent
2019-11-14 21:13:27
cassproject/cass-editor
https://api.github.com/repos/cassproject/cass-editor
closed
On attempt to export/publish some older frameworks get a TypeError on splice
Competency Framework High Priority bug
I was republishing the competency frameworks to fix the date formatting issue. Several frameworks failed to publish/export. If edit one of the frameworks (**Connecting Learning Theories to Your Teaching Strategies**), and then do a CE ASN JSON-LD export, I get the following error: TypeError: f["schema:publisher"].splice is not a function If I just edit, and save without making changes, I can publish without any errors. I left the above referenced framework as a reference, and fixed all of the other issue.
1.0
On attempt to export/publish some older frameworks get a TypeError on splice - I was republishing the competency frameworks to fix the date formatting issue. Several frameworks failed to publish/export. If edit one of the frameworks (**Connecting Learning Theories to Your Teaching Strategies**), and then do a CE ASN JSON-LD export, I get the following error: TypeError: f["schema:publisher"].splice is not a function If I just edit, and save without making changes, I can publish without any errors. I left the above referenced framework as a reference, and fixed all of the other issue.
priority
on attempt to export publish some older frameworks get a typeerror on splice i was republishing the competency frameworks to fix the date formatting issue several frameworks failed to publish export if edit one of the frameworks connecting learning theories to your teaching strategies and then do a ce asn json ld export i get the following error typeerror f splice is not a function if i just edit and save without making changes i can publish without any errors i left the above referenced framework as a reference and fixed all of the other issue
1
176,450
28,097,975,512
IssuesEvent
2023-03-30 17:10:24
Enterprise-CMCS/eAPD
https://api.github.com/repos/Enterprise-CMCS/eAPD
closed
[Design Issue] Create Closing Notes and Error Messages for Alt Considerations and Risks and Conditions for Enhanced Funding
Design small
The designs are done for these pages with the MES expansion, but we need closing notes, annotations, and error message designs for them as they transition to dev work ### This task is done when… - [ ] any acceptance criteria (not process oriented, requirements of feature) - [x] designs are created, taking into consideration validation, multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable - [x] 508 visual, content, and UX considerations and/or recommendations are reviewed - [x] designs are shared - [ ] designs are approved by product - [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer - [ ] a list of changes to be made if this is improving an existing feature --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
1.0
[Design Issue] Create Closing Notes and Error Messages for Alt Considerations and Risks and Conditions for Enhanced Funding - The designs are done for these pages with the MES expansion, but we need closing notes, annotations, and error message designs for them as they transition to dev work ### This task is done when… - [ ] any acceptance criteria (not process oriented, requirements of feature) - [x] designs are created, taking into consideration validation, multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable - [x] 508 visual, content, and UX considerations and/or recommendations are reviewed - [x] designs are shared - [ ] designs are approved by product - [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer - [ ] a list of changes to be made if this is improving an existing feature --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
non_priority
create closing notes and error messages for alt considerations and risks and conditions for enhanced funding the designs are done for these pages with the mes expansion but we need closing notes annotations and error message designs for them as they transition to dev work this task is done when… any acceptance criteria not process oriented requirements of feature designs are created taking into consideration validation multiple ffy multiple entries long entries weird entries and or the export view if applicable visual content and ux considerations and or recommendations are reviewed designs are shared designs are approved by product if there is a long discussion a summary of changes for dev will be added to the last comment before transfer a list of changes to be made if this is improving an existing feature add additional labels design dev compliance bug etc and size before submitting if the issue is needed to complete prioritized work for the current sprint add it to the this sprint pipeline otherwise all other issues will be automatically added to the unprioritized pipeline for prioritization in backlog refinement or sprint planning with product
0
476,946
13,752,818,441
IssuesEvent
2020-10-06 14:56:44
AY2021S1-CS2103T-W11-2/tp
https://api.github.com/repos/AY2021S1-CS2103T-W11-2/tp
closed
Add Tasks
priority.Medium severity.Low type.Story
As a user, I am able to add my personal tasks or tasks related to CS1101S but are not currently specified by any categories so that all CS1101S related tasks are stored together for easier reference. Tasks includes todos, deadlines and events.
1.0
Add Tasks - As a user, I am able to add my personal tasks or tasks related to CS1101S but are not currently specified by any categories so that all CS1101S related tasks are stored together for easier reference. Tasks includes todos, deadlines and events.
priority
add tasks as a user i am able to add my personal tasks or tasks related to but are not currently specified by any categories so that all related tasks are stored together for easier reference tasks includes todos deadlines and events
1
279,940
24,267,150,359
IssuesEvent
2022-09-28 07:22:39
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
SubViewports doesn't handle custom inputs
bug needs testing topic:input topic:gui
### Godot version Godot Engine v4.0.beta1.mono.official.4ba934bf3 ### System information Windows 11, Vulkan API 1.2.0 - Using Vulkan Device #0: AMD - AMD Radeon RX 5700 ### Issue description Using the _Input method on a viewport does not work. Recreating the same situation in Godot 3.5, using .Input and a viewport, it works as intended. Example of code, that should as far as I know send my current event to the SubViewport ``` public override void _Input(InputEvent @event) { if (@event is InputEventMouse _mouseEvent) { (GetNode<Viewport>("SubViewport") as Viewport)._Input(@event); } } ``` ### Steps to reproduce Create a viewport which includes any element which accept an input, and try to give the viewport an input through code. ### Minimal reproduction project [InputExample.zip](https://github.com/godotengine/godot/files/9648710/InputExample.zip)
1.0
SubViewports doesn't handle custom inputs - ### Godot version Godot Engine v4.0.beta1.mono.official.4ba934bf3 ### System information Windows 11, Vulkan API 1.2.0 - Using Vulkan Device #0: AMD - AMD Radeon RX 5700 ### Issue description Using the _Input method on a viewport does not work. Recreating the same situation in Godot 3.5, using .Input and a viewport, it works as intended. Example of code, that should as far as I know send my current event to the SubViewport ``` public override void _Input(InputEvent @event) { if (@event is InputEventMouse _mouseEvent) { (GetNode<Viewport>("SubViewport") as Viewport)._Input(@event); } } ``` ### Steps to reproduce Create a viewport which includes any element which accept an input, and try to give the viewport an input through code. ### Minimal reproduction project [InputExample.zip](https://github.com/godotengine/godot/files/9648710/InputExample.zip)
non_priority
subviewports doesn t handle custom inputs godot version godot engine mono official system information windows vulkan api using vulkan device amd amd radeon rx issue description using the input method on a viewport does not work recreating the same situation in godot using input and a viewport it works as intended example of code that should as far as i know send my current event to the subviewport public override void input inputevent event if event is inputeventmouse mouseevent getnode subviewport as viewport input event steps to reproduce create a viewport which includes any element which accept an input and try to give the viewport an input through code minimal reproduction project
0
656,742
21,774,071,895
IssuesEvent
2022-05-13 12:07:22
metabase/metabase
https://api.github.com/repos/metabase/metabase
opened
Custom Expression aggregation in Metrics does not work
Type:Bug Priority:P1 .Correctness Administration/Metrics & Segments .Frontend .Regression
**Describe the bug** Custom Expression aggregation in Metrics does not work. Regression since 0.43.0 **To Reproduce** 1. Admin > Data Model > Metrics > Create > Sample > Orders 2. Click "View" > Custom Expression 1. The modal is too small 2. There is no formula validation, so incorrect things can be input - ex. `Count / Distinct([BLAHBLAHBLAH])` ![image](https://user-images.githubusercontent.com/1447303/168277390-fc1028ec-b6ef-4f88-ac73-27218bcf7117.png) 3. Inputting something like `Count / Distinct([Product ID])` and clicking Done ![image](https://user-images.githubusercontent.com/1447303/168278550-645474f5-57a9-4e0a-93c0-b38485da72b5.png) Will look like it's saving the expression, but the results are incorrect (should be `93.8`) and opening the expression again will not show the formula that was inputted ![image](https://user-images.githubusercontent.com/1447303/168278861-5ed537fb-7dfd-4165-993c-39a26ca8e8cb.png) **Expected behavior** Similar to Custom Expression aggregation in normal Notebook **Information about your Metabase Installation:** Tested 0.42.4 thru 0.43.0 - regression since 0.43.0
1.0
Custom Expression aggregation in Metrics does not work - **Describe the bug** Custom Expression aggregation in Metrics does not work. Regression since 0.43.0 **To Reproduce** 1. Admin > Data Model > Metrics > Create > Sample > Orders 2. Click "View" > Custom Expression 1. The modal is too small 2. There is no formula validation, so incorrect things can be input - ex. `Count / Distinct([BLAHBLAHBLAH])` ![image](https://user-images.githubusercontent.com/1447303/168277390-fc1028ec-b6ef-4f88-ac73-27218bcf7117.png) 3. Inputting something like `Count / Distinct([Product ID])` and clicking Done ![image](https://user-images.githubusercontent.com/1447303/168278550-645474f5-57a9-4e0a-93c0-b38485da72b5.png) Will look like it's saving the expression, but the results are incorrect (should be `93.8`) and opening the expression again will not show the formula that was inputted ![image](https://user-images.githubusercontent.com/1447303/168278861-5ed537fb-7dfd-4165-993c-39a26ca8e8cb.png) **Expected behavior** Similar to Custom Expression aggregation in normal Notebook **Information about your Metabase Installation:** Tested 0.42.4 thru 0.43.0 - regression since 0.43.0
priority
custom expression aggregation in metrics does not work describe the bug custom expression aggregation in metrics does not work regression since to reproduce admin data model metrics create sample orders click view custom expression the modal is too small there is no formula validation so incorrect things can be input ex count distinct inputting something like count distinct and clicking done will look like it s saving the expression but the results are incorrect should be and opening the expression again will not show the formula that was inputted expected behavior similar to custom expression aggregation in normal notebook information about your metabase installation tested thru regression since
1
130,304
27,642,403,633
IssuesEvent
2023-03-10 19:17:19
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
opened
Code Table Request - UCR: University of California Riverside Herbarium
Function-CodeTables
## Instructions This is a template to facilitate communication with the Arctos Code Table Committee. Submit a separate request for each relevant value. This form is appropriate for exploring how data may best be stored, for adding vocabulary, or for updating existing definitions. Reviewing documentation before proceeding will result in a more enjoyable experience. * [Issue Documentation](http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html) * [Code Table Documentation](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) * [Video Tutorial - Submit a Code Table Request](https://youtu.be/t2jHbsRA3lk) ------------------------------ ## Initial Request **Goal**: Describe what you're trying to accomplish. This is the only necessary step to start this process. The Committee is available to assist with all other steps. Please clearly indicate any uncertainty or desired guidance if you proceed beyond this step. All [UCR: University of California Riverside Herbarium](https://arctos.database.museum/info/ctDocumentation.cfm?table=ctcoll_other_id_type#ucr__university_of_california_riverside_herbarium) should be replaced with other ID type = other identifier and issued by agent [University of California, Riverside Herbarium](https://arctos.database.museum/agent/21347654) **Proposed Value**: Proposed new value. This should be clear and compatible with similar values in the relevant table and across Arctos. **Proposed Definition**: Clear, complete, non-collection-type-specific **functional** definition of the value. Avoid discipline-specific terminology if possible, include parenthetically if unavoidable. **Context**: Describe why this new value is necessary and existing values are not. **Table**: Code Tables are http://arctos.database.museum/info/ctDocumentation.cfm. Link to the specific table or value. This may involve multiple tables and will control datatype for Attributes. OtherID requests require BaseURL (and example) or explanation. Please ask for assistance if unsure. **Collection type**: Some code tables contain collection-type-specific values. ``collection_cde`` may be found from https://arctos.database.museum/home.cfm **Priority**: Please describe the urgency and/or choose a priority-label to the right. You should expect a response within two working days, and may utilize [Arctos Contacts](https://arctosdb.org/contacts/) if you feel response is lacking. **Available for Public View**: Most data are by default publicly available. Describe any necessary access restrictions. **Project**: Add the issue to the [Code Table Management Project](https://github.com/ArctosDB/arctos/projects/13#card-31628184). **Discussion**: Please reach out to anyone who might be affected by this change. Leave a comment or add this to the Committee agenda if you believe more focused conversation is necessary. ## Approval All of the following must be checked before this may proceed. The [How-To Document](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) should be followed. Pay particular attention to terminology (with emphasis on consistency) and documentation (with emphasis on functionality). - [ ] Code Table Administrator[1] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] Code Table Administrator[2] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] DBA - The request is functionally acceptable. The term is not a functional duplicate, and is compatible with existing data and code. - [ ] DBA - Appropriate code or handlers are in place as necessary. (ID_References, Media Relationships, Encumbrances, etc. require particular attention) ## Rejection If you believe this request should not proceed, explain why here. Suggest any changes that would make the change acceptable, alternate (usually existing) paths to the same goals, etc. 1. Can a suitable solution be found here? If not, proceed to (2) 2. Can a suitable solution be found by Code Table Committee discussion? If not, proceed to (3) 3. Take the discussion to a monthly Arctos Working Group meeting for final resolution. ## Implementation Once all of the Approval Checklist is appropriately checked and there are no Rejection comments, or in special circumstances by decree of the Arctos Working Group, the change may be made. Review everything one last time. Ensure the How-To has been followed. Ensure all checks have been made by appropriate personnel. Make changes as described above. Ensure the URL of this Issue is included in the definition. Close this Issue. **DO NOT** modify Arctos Authorities in any way before all points in this Issue have been fully addressed; data loss may result. ## Special Exemptions In very specific cases and by prior approval of The Committee, the approval process may be skipped, and implementation requirements may be slightly altered. Please note here if you are proceeding under one of these use cases. 1. Adding an existing term to additional collection types may proceed immediately and without discussion, but doing so may also subject users to future cleanup efforts. If time allows, please review the term and definition as part of this step. 2. The Committee may grant special access on particular tables to particular users. This should be exercised with great caution only after several smooth test cases, and generally limited to "taxonomy-like" data such as International Commission on Stratigraphy terminology.
1.0
Code Table Request - UCR: University of California Riverside Herbarium - ## Instructions This is a template to facilitate communication with the Arctos Code Table Committee. Submit a separate request for each relevant value. This form is appropriate for exploring how data may best be stored, for adding vocabulary, or for updating existing definitions. Reviewing documentation before proceeding will result in a more enjoyable experience. * [Issue Documentation](http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html) * [Code Table Documentation](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) * [Video Tutorial - Submit a Code Table Request](https://youtu.be/t2jHbsRA3lk) ------------------------------ ## Initial Request **Goal**: Describe what you're trying to accomplish. This is the only necessary step to start this process. The Committee is available to assist with all other steps. Please clearly indicate any uncertainty or desired guidance if you proceed beyond this step. All [UCR: University of California Riverside Herbarium](https://arctos.database.museum/info/ctDocumentation.cfm?table=ctcoll_other_id_type#ucr__university_of_california_riverside_herbarium) should be replaced with other ID type = other identifier and issued by agent [University of California, Riverside Herbarium](https://arctos.database.museum/agent/21347654) **Proposed Value**: Proposed new value. This should be clear and compatible with similar values in the relevant table and across Arctos. **Proposed Definition**: Clear, complete, non-collection-type-specific **functional** definition of the value. Avoid discipline-specific terminology if possible, include parenthetically if unavoidable. **Context**: Describe why this new value is necessary and existing values are not. **Table**: Code Tables are http://arctos.database.museum/info/ctDocumentation.cfm. Link to the specific table or value. This may involve multiple tables and will control datatype for Attributes. OtherID requests require BaseURL (and example) or explanation. Please ask for assistance if unsure. **Collection type**: Some code tables contain collection-type-specific values. ``collection_cde`` may be found from https://arctos.database.museum/home.cfm **Priority**: Please describe the urgency and/or choose a priority-label to the right. You should expect a response within two working days, and may utilize [Arctos Contacts](https://arctosdb.org/contacts/) if you feel response is lacking. **Available for Public View**: Most data are by default publicly available. Describe any necessary access restrictions. **Project**: Add the issue to the [Code Table Management Project](https://github.com/ArctosDB/arctos/projects/13#card-31628184). **Discussion**: Please reach out to anyone who might be affected by this change. Leave a comment or add this to the Committee agenda if you believe more focused conversation is necessary. ## Approval All of the following must be checked before this may proceed. The [How-To Document](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) should be followed. Pay particular attention to terminology (with emphasis on consistency) and documentation (with emphasis on functionality). - [ ] Code Table Administrator[1] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] Code Table Administrator[2] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] DBA - The request is functionally acceptable. The term is not a functional duplicate, and is compatible with existing data and code. - [ ] DBA - Appropriate code or handlers are in place as necessary. (ID_References, Media Relationships, Encumbrances, etc. require particular attention) ## Rejection If you believe this request should not proceed, explain why here. Suggest any changes that would make the change acceptable, alternate (usually existing) paths to the same goals, etc. 1. Can a suitable solution be found here? If not, proceed to (2) 2. Can a suitable solution be found by Code Table Committee discussion? If not, proceed to (3) 3. Take the discussion to a monthly Arctos Working Group meeting for final resolution. ## Implementation Once all of the Approval Checklist is appropriately checked and there are no Rejection comments, or in special circumstances by decree of the Arctos Working Group, the change may be made. Review everything one last time. Ensure the How-To has been followed. Ensure all checks have been made by appropriate personnel. Make changes as described above. Ensure the URL of this Issue is included in the definition. Close this Issue. **DO NOT** modify Arctos Authorities in any way before all points in this Issue have been fully addressed; data loss may result. ## Special Exemptions In very specific cases and by prior approval of The Committee, the approval process may be skipped, and implementation requirements may be slightly altered. Please note here if you are proceeding under one of these use cases. 1. Adding an existing term to additional collection types may proceed immediately and without discussion, but doing so may also subject users to future cleanup efforts. If time allows, please review the term and definition as part of this step. 2. The Committee may grant special access on particular tables to particular users. This should be exercised with great caution only after several smooth test cases, and generally limited to "taxonomy-like" data such as International Commission on Stratigraphy terminology.
non_priority
code table request ucr university of california riverside herbarium instructions this is a template to facilitate communication with the arctos code table committee submit a separate request for each relevant value this form is appropriate for exploring how data may best be stored for adding vocabulary or for updating existing definitions reviewing documentation before proceeding will result in a more enjoyable experience initial request goal describe what you re trying to accomplish this is the only necessary step to start this process the committee is available to assist with all other steps please clearly indicate any uncertainty or desired guidance if you proceed beyond this step all should be replaced with other id type other identifier and issued by agent proposed value proposed new value this should be clear and compatible with similar values in the relevant table and across arctos proposed definition clear complete non collection type specific functional definition of the value avoid discipline specific terminology if possible include parenthetically if unavoidable context describe why this new value is necessary and existing values are not table code tables are link to the specific table or value this may involve multiple tables and will control datatype for attributes otherid requests require baseurl and example or explanation please ask for assistance if unsure collection type some code tables contain collection type specific values collection cde may be found from priority please describe the urgency and or choose a priority label to the right you should expect a response within two working days and may utilize if you feel response is lacking available for public view most data are by default publicly available describe any necessary access restrictions project add the issue to the discussion please reach out to anyone who might be affected by this change leave a comment or add this to the committee agenda if you believe more focused conversation is necessary approval all of the following must be checked before this may proceed the should be followed pay particular attention to terminology with emphasis on consistency and documentation with emphasis on functionality code table administrator check and initial comment or thumbs up to indicate that the request complies with the how to documentation and has your approval code table administrator check and initial comment or thumbs up to indicate that the request complies with the how to documentation and has your approval dba the request is functionally acceptable the term is not a functional duplicate and is compatible with existing data and code dba appropriate code or handlers are in place as necessary id references media relationships encumbrances etc require particular attention rejection if you believe this request should not proceed explain why here suggest any changes that would make the change acceptable alternate usually existing paths to the same goals etc can a suitable solution be found here if not proceed to can a suitable solution be found by code table committee discussion if not proceed to take the discussion to a monthly arctos working group meeting for final resolution implementation once all of the approval checklist is appropriately checked and there are no rejection comments or in special circumstances by decree of the arctos working group the change may be made review everything one last time ensure the how to has been followed ensure all checks have been made by appropriate personnel make changes as described above ensure the url of this issue is included in the definition close this issue do not modify arctos authorities in any way before all points in this issue have been fully addressed data loss may result special exemptions in very specific cases and by prior approval of the committee the approval process may be skipped and implementation requirements may be slightly altered please note here if you are proceeding under one of these use cases adding an existing term to additional collection types may proceed immediately and without discussion but doing so may also subject users to future cleanup efforts if time allows please review the term and definition as part of this step the committee may grant special access on particular tables to particular users this should be exercised with great caution only after several smooth test cases and generally limited to taxonomy like data such as international commission on stratigraphy terminology
0
305,972
9,379,097,375
IssuesEvent
2019-04-04 14:16:28
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
reopened
ZONE_NAME header gets stripped by passenger
Category: Drivers - Network Category: Sunstone Priority: High Sponsored Status: Accepted Type: Bug
**Description** ZONE_NAME header gets stripped by passenger **To Reproduce** change zone using sunstone + passenger **Expected behavior** zone should be changed **Details** - Affected Component: Sunstone <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
ZONE_NAME header gets stripped by passenger - **Description** ZONE_NAME header gets stripped by passenger **To Reproduce** change zone using sunstone + passenger **Expected behavior** zone should be changed **Details** - Affected Component: Sunstone <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
priority
zone name header gets stripped by passenger description zone name header gets stripped by passenger to reproduce change zone using sunstone passenger expected behavior zone should be changed details affected component sunstone progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
1
14,625
3,412,316,260
IssuesEvent
2015-12-05 20:01:21
radare/radare2
https://api.github.com/repos/radare/radare2
closed
ELF: no imported functions identification
bug file-format test-required
On this binary: https://github.com/ctfs/write-ups-2015/tree/master/nuit-du-hack-ctf-quals-2015/reverse/clark-kent r2 fails to correctly identify imported functions. This is probably because the binary doesn't have section header information.
1.0
ELF: no imported functions identification - On this binary: https://github.com/ctfs/write-ups-2015/tree/master/nuit-du-hack-ctf-quals-2015/reverse/clark-kent r2 fails to correctly identify imported functions. This is probably because the binary doesn't have section header information.
non_priority
elf no imported functions identification on this binary fails to correctly identify imported functions this is probably because the binary doesn t have section header information
0
535,501
15,689,435,358
IssuesEvent
2021-03-25 15:40:29
fecgov/fec-eregs
https://api.github.com/repos/fecgov/fec-eregs
opened
Load the 2021 11 CFR into e-regs
High priority Pairing opportunity Work: Back-end Work: Content
### What we are after Making sure the newest version (2021) of 11 CFR is parsed into e-regs GPO links to different formats: will add link here when available Content Team will put in changes in for the new regs so we know what to look for. ### Completion criteria - [ ] Regs loaded onto dev - [ ] Regs loaded onto stage - [ ] New regulations are loaded to the live website (Regs loaded onto prod) ### Tech considerations - [ ] Reload the regulations in dev and stage - [ ] Test in dev and stage with content team - [ ] Reload regulations in prod - [ ] Have team verify the regulations are updated in production
1.0
Load the 2021 11 CFR into e-regs - ### What we are after Making sure the newest version (2021) of 11 CFR is parsed into e-regs GPO links to different formats: will add link here when available Content Team will put in changes in for the new regs so we know what to look for. ### Completion criteria - [ ] Regs loaded onto dev - [ ] Regs loaded onto stage - [ ] New regulations are loaded to the live website (Regs loaded onto prod) ### Tech considerations - [ ] Reload the regulations in dev and stage - [ ] Test in dev and stage with content team - [ ] Reload regulations in prod - [ ] Have team verify the regulations are updated in production
priority
load the cfr into e regs what we are after making sure the newest version of cfr is parsed into e regs gpo links to different formats will add link here when available content team will put in changes in for the new regs so we know what to look for completion criteria regs loaded onto dev regs loaded onto stage new regulations are loaded to the live website regs loaded onto prod tech considerations reload the regulations in dev and stage test in dev and stage with content team reload regulations in prod have team verify the regulations are updated in production
1
42,122
9,162,056,792
IssuesEvent
2019-03-01 12:18:04
mozilla-releng/firefox-infra-changelog
https://api.github.com/repos/mozilla-releng/firefox-infra-changelog
closed
W0622: Redefining built-in 'all' (redefined-builtin) in file 'client.py'
code-style
```Python client.py:30:8: W0622: Redefining built-in 'all' (redefined-builtin) ``` We should rename the variable to use something else than `all`. Background: all() is a method that returns a bool True value when all elements in the given iterable are true.
1.0
W0622: Redefining built-in 'all' (redefined-builtin) in file 'client.py' - ```Python client.py:30:8: W0622: Redefining built-in 'all' (redefined-builtin) ``` We should rename the variable to use something else than `all`. Background: all() is a method that returns a bool True value when all elements in the given iterable are true.
non_priority
redefining built in all redefined builtin in file client py python client py redefining built in all redefined builtin we should rename the variable to use something else than all background all is a method that returns a bool true value when all elements in the given iterable are true
0
240,094
20,010,512,320
IssuesEvent
2022-02-01 05:28:33
open-telemetry/opentelemetry-java-instrumentation
https://api.github.com/repos/open-telemetry/opentelemetry-java-instrumentation
opened
Run tests against more versions in muzzle range
enhancement area:tests
Currently tests are typically run against the initial (base) and latest version in the muzzle range. The muzzle checks tell us that there are no API-incompatible changes in the versions in the middle that affect the instrumentation. But what muzzle can't tell us is if there were any behavior changes in the versions in the middle that could somehow affect things, e.g. #5027 where the API signature didn't change, but the behavior of the API changed. This happens fairly rarely, but it still may be worth testing against more versions to help close this gap. A reasonable option seems to be to test against all initial major versions (or some configurable definition of versions), e.g. hibernate 5.0.0 (#5210), when the muzzle range spans multiple major versions (or some long time frame of releases).
1.0
Run tests against more versions in muzzle range - Currently tests are typically run against the initial (base) and latest version in the muzzle range. The muzzle checks tell us that there are no API-incompatible changes in the versions in the middle that affect the instrumentation. But what muzzle can't tell us is if there were any behavior changes in the versions in the middle that could somehow affect things, e.g. #5027 where the API signature didn't change, but the behavior of the API changed. This happens fairly rarely, but it still may be worth testing against more versions to help close this gap. A reasonable option seems to be to test against all initial major versions (or some configurable definition of versions), e.g. hibernate 5.0.0 (#5210), when the muzzle range spans multiple major versions (or some long time frame of releases).
non_priority
run tests against more versions in muzzle range currently tests are typically run against the initial base and latest version in the muzzle range the muzzle checks tell us that there are no api incompatible changes in the versions in the middle that affect the instrumentation but what muzzle can t tell us is if there were any behavior changes in the versions in the middle that could somehow affect things e g where the api signature didn t change but the behavior of the api changed this happens fairly rarely but it still may be worth testing against more versions to help close this gap a reasonable option seems to be to test against all initial major versions or some configurable definition of versions e g hibernate when the muzzle range spans multiple major versions or some long time frame of releases
0
375,836
11,135,042,653
IssuesEvent
2019-12-20 13:27:08
Rithari/OnePlusBot
https://api.github.com/repos/Rithari/OnePlusBot
opened
Rework the way modmail messages are being logged
enhancement low priority
**Is your feature request related to a problem? Please describe.** Currently, when closing, each embed is replayed into the logging channel, which, because for ratelimiting reasons, is delayed. This mechanism causes the logging to take a significant amount. **Describe the solution you'd like** Logging mechanism should be changed, so that less messages are needed for logging and the closing of a thread can happen much faster. This might happen via an attached file to an embed. *Additionally* the close mechanism needs to check whether or not the thread already is getting closed.
1.0
Rework the way modmail messages are being logged - **Is your feature request related to a problem? Please describe.** Currently, when closing, each embed is replayed into the logging channel, which, because for ratelimiting reasons, is delayed. This mechanism causes the logging to take a significant amount. **Describe the solution you'd like** Logging mechanism should be changed, so that less messages are needed for logging and the closing of a thread can happen much faster. This might happen via an attached file to an embed. *Additionally* the close mechanism needs to check whether or not the thread already is getting closed.
priority
rework the way modmail messages are being logged is your feature request related to a problem please describe currently when closing each embed is replayed into the logging channel which because for ratelimiting reasons is delayed this mechanism causes the logging to take a significant amount describe the solution you d like logging mechanism should be changed so that less messages are needed for logging and the closing of a thread can happen much faster this might happen via an attached file to an embed additionally the close mechanism needs to check whether or not the thread already is getting closed
1
173,733
6,530,405,298
IssuesEvent
2017-08-30 15:01:04
axvr/Vivid.vim
https://api.github.com/repos/axvr/Vivid.vim
reopened
Modify plugin install location
enhancement priority
Automatic setting of plugin install location depending upon running Vim or Neovim. Allow user to set a custom location.
1.0
Modify plugin install location - Automatic setting of plugin install location depending upon running Vim or Neovim. Allow user to set a custom location.
priority
modify plugin install location automatic setting of plugin install location depending upon running vim or neovim allow user to set a custom location
1
690,494
23,661,759,237
IssuesEvent
2022-08-26 16:15:23
o3de/o3de
https://api.github.com/repos/o3de/o3de
closed
Lua Editor: Unable to connect to the O3DE Editor
feature/networking kind/bug sig/content triage/accepted priority/critical
**Describe the bug** The Lua Editor cannot be connected to the O3DE Editor. On the first launch of the Lua Editor, the only available options in Target list are AssetProcessor and AssetBuilder. When the Lua Editor is restarted, these are no longer available and the Target list is empty. Additionally, when the Lua Editor is closed for the first time, the following Assert appears in the Console: ``` Trace::Assert D:\Git\o3de\Code\Framework\AzCore\AzCore/std/parallel/thread.h(99): (12188) 'class AZStd::thread &__cdecl AZStd::thread::operator =(class AZStd::thread &&)' You must call detach or join before you delete/move over the current thread! ``` Please refer to the attached video and assert.txt file for more details. **Steps to reproduce** Steps to reproduce the behavior: 1. Launch O3DE Editor. 2. Launch Lua Editor (Tools -> Lua Editor). 3. Switch focus to the O3DE Editor (due to #7237). 4. Switch focus back to the Lua Editor. 5. Click the Target button to attempt to connect to the O3DE Editor. 6. Restart the LUA editor. 7. Repeat steps 3-6. **Expected behavior** The O3DE Editor appears on the Target list and can be connected to. **Actual behavior** The O3DE Editor does not appear on the Target list and cannot be connected to. **Screenshots/Video** https://user-images.githubusercontent.com/86953108/165092336-29204b7d-1317-45eb-a8c4-d05e32263236.mp4 **Found in Branch** Development (f590593) **Desktop/Device:** - Device: PC - OS: Windows - Version 10 - CPU AMD Ryzen 5 3600 - GPU Nvidia RTX 2060 Super - Memory 16GB **Attachments** [Assert.txt](https://github.com/o3de/o3de/files/8554668/Assert.txt)
1.0
Lua Editor: Unable to connect to the O3DE Editor - **Describe the bug** The Lua Editor cannot be connected to the O3DE Editor. On the first launch of the Lua Editor, the only available options in Target list are AssetProcessor and AssetBuilder. When the Lua Editor is restarted, these are no longer available and the Target list is empty. Additionally, when the Lua Editor is closed for the first time, the following Assert appears in the Console: ``` Trace::Assert D:\Git\o3de\Code\Framework\AzCore\AzCore/std/parallel/thread.h(99): (12188) 'class AZStd::thread &__cdecl AZStd::thread::operator =(class AZStd::thread &&)' You must call detach or join before you delete/move over the current thread! ``` Please refer to the attached video and assert.txt file for more details. **Steps to reproduce** Steps to reproduce the behavior: 1. Launch O3DE Editor. 2. Launch Lua Editor (Tools -> Lua Editor). 3. Switch focus to the O3DE Editor (due to #7237). 4. Switch focus back to the Lua Editor. 5. Click the Target button to attempt to connect to the O3DE Editor. 6. Restart the LUA editor. 7. Repeat steps 3-6. **Expected behavior** The O3DE Editor appears on the Target list and can be connected to. **Actual behavior** The O3DE Editor does not appear on the Target list and cannot be connected to. **Screenshots/Video** https://user-images.githubusercontent.com/86953108/165092336-29204b7d-1317-45eb-a8c4-d05e32263236.mp4 **Found in Branch** Development (f590593) **Desktop/Device:** - Device: PC - OS: Windows - Version 10 - CPU AMD Ryzen 5 3600 - GPU Nvidia RTX 2060 Super - Memory 16GB **Attachments** [Assert.txt](https://github.com/o3de/o3de/files/8554668/Assert.txt)
priority
lua editor unable to connect to the editor describe the bug the lua editor cannot be connected to the editor on the first launch of the lua editor the only available options in target list are assetprocessor and assetbuilder when the lua editor is restarted these are no longer available and the target list is empty additionally when the lua editor is closed for the first time the following assert appears in the console trace assert d git code framework azcore azcore std parallel thread h class azstd thread cdecl azstd thread operator class azstd thread you must call detach or join before you delete move over the current thread please refer to the attached video and assert txt file for more details steps to reproduce steps to reproduce the behavior launch editor launch lua editor tools lua editor switch focus to the editor due to switch focus back to the lua editor click the target button to attempt to connect to the editor restart the lua editor repeat steps expected behavior the editor appears on the target list and can be connected to actual behavior the editor does not appear on the target list and cannot be connected to screenshots video found in branch development desktop device device pc os windows version cpu amd ryzen gpu nvidia rtx super memory attachments
1
101,546
21,709,364,207
IssuesEvent
2022-05-10 12:39:05
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: Allow query/js confirmation modal to be dismissed by clicking outside
Bug UX Improvement Production Needs Triaging Query Execution BE Coders Pod medium Actions Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Currently, user has to click on `yes` or `no` buttons on the confirmation modal to close the modal - but clicking outside the modal should also be allowed when user wants to quickly dismiss the action and continue on. ### Steps To Reproduce 1. Mark any query or JS function to request user confirmation and run the respective query/function. Observe that quick dismiss by clicking outside is not allowed ### Public Sample App [Sample app](https://release.app.appsmith.com/applications/62385770e2e3e942f7cdb921/pages/62385770e2e3e942f7cdb925/edit) ### Version Cloud
1.0
[Bug]: Allow query/js confirmation modal to be dismissed by clicking outside - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Currently, user has to click on `yes` or `no` buttons on the confirmation modal to close the modal - but clicking outside the modal should also be allowed when user wants to quickly dismiss the action and continue on. ### Steps To Reproduce 1. Mark any query or JS function to request user confirmation and run the respective query/function. Observe that quick dismiss by clicking outside is not allowed ### Public Sample App [Sample app](https://release.app.appsmith.com/applications/62385770e2e3e942f7cdb921/pages/62385770e2e3e942f7cdb925/edit) ### Version Cloud
non_priority
allow query js confirmation modal to be dismissed by clicking outside is there an existing issue for this i have searched the existing issues description currently user has to click on yes or no buttons on the confirmation modal to close the modal but clicking outside the modal should also be allowed when user wants to quickly dismiss the action and continue on steps to reproduce mark any query or js function to request user confirmation and run the respective query function observe that quick dismiss by clicking outside is not allowed public sample app version cloud
0
83,435
24,055,870,278
IssuesEvent
2022-09-16 16:48:49
gradle/gradle
https://api.github.com/repos/gradle/gradle
closed
Improve 'dependency not found' error message to be less misleading about the cause of the problem
a:feature in:composite-builds stale
@adammurdoch commented on [Thu Jun 15 2017](https://github.com/gradle/composite-builds/issues/116) The error messages that Gradle produces when a match for a given selector cannot be found all assume that the only place we look is in repositories defined by the project. For example, we have error messages that say things like "searched in these locations: <some-urls>" or "no repositories defined" or "we're in offline mode" (these aren't literally the error messages, just examples). However, it is no longer the case that the only place we search is in repositories, and there are also legit use cases where I don't have any repositories defined. If, for example, I make a typo in the build script for the consuming build, then i get told "you don't have any repositories defined" when in fact this is not the problem at all and is misleading as to what I should do next. The error messages should at the very least be worded to reflect the fact that we also looked for matches in all the builds in the current Gradle invocation (including the current build), if any. Even better if the message gave some details of what might have matched (assuming a typo) or provides some hint to discover what is available for matching, e.g. "please run this report" or whatever.
1.0
Improve 'dependency not found' error message to be less misleading about the cause of the problem - @adammurdoch commented on [Thu Jun 15 2017](https://github.com/gradle/composite-builds/issues/116) The error messages that Gradle produces when a match for a given selector cannot be found all assume that the only place we look is in repositories defined by the project. For example, we have error messages that say things like "searched in these locations: <some-urls>" or "no repositories defined" or "we're in offline mode" (these aren't literally the error messages, just examples). However, it is no longer the case that the only place we search is in repositories, and there are also legit use cases where I don't have any repositories defined. If, for example, I make a typo in the build script for the consuming build, then i get told "you don't have any repositories defined" when in fact this is not the problem at all and is misleading as to what I should do next. The error messages should at the very least be worded to reflect the fact that we also looked for matches in all the builds in the current Gradle invocation (including the current build), if any. Even better if the message gave some details of what might have matched (assuming a typo) or provides some hint to discover what is available for matching, e.g. "please run this report" or whatever.
non_priority
improve dependency not found error message to be less misleading about the cause of the problem adammurdoch commented on the error messages that gradle produces when a match for a given selector cannot be found all assume that the only place we look is in repositories defined by the project for example we have error messages that say things like searched in these locations or no repositories defined or we re in offline mode these aren t literally the error messages just examples however it is no longer the case that the only place we search is in repositories and there are also legit use cases where i don t have any repositories defined if for example i make a typo in the build script for the consuming build then i get told you don t have any repositories defined when in fact this is not the problem at all and is misleading as to what i should do next the error messages should at the very least be worded to reflect the fact that we also looked for matches in all the builds in the current gradle invocation including the current build if any even better if the message gave some details of what might have matched assuming a typo or provides some hint to discover what is available for matching e g please run this report or whatever
0
119,575
15,574,721,597
IssuesEvent
2021-03-17 10:11:02
OfficeDev/office-js
https://api.github.com/repos/OfficeDev/office-js
closed
In Excel Online, the automatically created worksheet of all new workbooks, have the same id
Area: Excel Resolution: by design Type: product bug
As described in [this Stack Overflow question](https://stackoverflow.com/questions/56502276/all-first-worksheets-of-files-created-in-excel-online-have-the-same-id), when I go to [office.com](https://www.office.com) and create a new Excel file, the worksheet that already exists in the new file, has always id `{00000000-0001-0000-0000-000000000000}`. If I manually create a new worksheet, the worksheet has a normal-looking random GUID id. So it seems that there's just an issue with the automatically created worksheet of new files. Note that when using the latest Excel (desktop) for macOS connected to Office 365, this issue does not exist. First worksheets of new Excel files have random GUID ids. ## Expected Behavior The automatically created worksheet of new Excel Online files should have a random GUID id. ## Current Behavior The initial worksheet of all new Excel online files has always id `{00000000-0001-0000-0000-000000000000}`. ## Steps to Reproduce, or Live Example This Script Lab snippet just prints the current worksheet id to the console: https://gist.github.com/tiagoalves/6afaaeb014e4ebdc9392e64c8e6246a3 Go to https://www.office.com, create a new Excel file and run the snippet without creating any new worksheets. See the screen recording linked at the end for a demo. ## Context I'm creating an Excel Add-in that associates certain information of the add-in to a worksheet and I'm using the worksheet id as a unique identifier. However, with this apparent issue, with worksheet ids that collide, I can't rely on the worksheet id. ## Your Environment * Platform [PC desktop, Mac, iOS, Office Online]: Office Online * Host [Excel, Word, PowerPoint, etc.]: Excel * Office version number: latest? * Operating System: macOS 10.14.4 * Browser (if using Office Online): Firefox 67.0.1, Chrome 74.0.3729.169. ## Useful logs This screen recording shows the issue being reproduced: https://imgur.com/a/pdYbtk5
1.0
In Excel Online, the automatically created worksheet of all new workbooks, have the same id - As described in [this Stack Overflow question](https://stackoverflow.com/questions/56502276/all-first-worksheets-of-files-created-in-excel-online-have-the-same-id), when I go to [office.com](https://www.office.com) and create a new Excel file, the worksheet that already exists in the new file, has always id `{00000000-0001-0000-0000-000000000000}`. If I manually create a new worksheet, the worksheet has a normal-looking random GUID id. So it seems that there's just an issue with the automatically created worksheet of new files. Note that when using the latest Excel (desktop) for macOS connected to Office 365, this issue does not exist. First worksheets of new Excel files have random GUID ids. ## Expected Behavior The automatically created worksheet of new Excel Online files should have a random GUID id. ## Current Behavior The initial worksheet of all new Excel online files has always id `{00000000-0001-0000-0000-000000000000}`. ## Steps to Reproduce, or Live Example This Script Lab snippet just prints the current worksheet id to the console: https://gist.github.com/tiagoalves/6afaaeb014e4ebdc9392e64c8e6246a3 Go to https://www.office.com, create a new Excel file and run the snippet without creating any new worksheets. See the screen recording linked at the end for a demo. ## Context I'm creating an Excel Add-in that associates certain information of the add-in to a worksheet and I'm using the worksheet id as a unique identifier. However, with this apparent issue, with worksheet ids that collide, I can't rely on the worksheet id. ## Your Environment * Platform [PC desktop, Mac, iOS, Office Online]: Office Online * Host [Excel, Word, PowerPoint, etc.]: Excel * Office version number: latest? * Operating System: macOS 10.14.4 * Browser (if using Office Online): Firefox 67.0.1, Chrome 74.0.3729.169. ## Useful logs This screen recording shows the issue being reproduced: https://imgur.com/a/pdYbtk5
non_priority
in excel online the automatically created worksheet of all new workbooks have the same id as described in when i go to and create a new excel file the worksheet that already exists in the new file has always id if i manually create a new worksheet the worksheet has a normal looking random guid id so it seems that there s just an issue with the automatically created worksheet of new files note that when using the latest excel desktop for macos connected to office this issue does not exist first worksheets of new excel files have random guid ids expected behavior the automatically created worksheet of new excel online files should have a random guid id current behavior the initial worksheet of all new excel online files has always id steps to reproduce or live example this script lab snippet just prints the current worksheet id to the console go to create a new excel file and run the snippet without creating any new worksheets see the screen recording linked at the end for a demo context i m creating an excel add in that associates certain information of the add in to a worksheet and i m using the worksheet id as a unique identifier however with this apparent issue with worksheet ids that collide i can t rely on the worksheet id your environment platform office online host excel office version number latest operating system macos browser if using office online firefox chrome useful logs this screen recording shows the issue being reproduced
0
371,585
10,974,366,452
IssuesEvent
2019-11-29 08:58:13
deepnetworkgmbh/deepnetworkgmbh.github.io
https://api.github.com/repos/deepnetworkgmbh/deepnetworkgmbh.github.io
closed
Garbled UI Elements on Chrome/Windows on first load.
bug high priority
**Repro steps** - either ctrl-F5, or do a first time load of home page. - Issue resolves after first slider action. Expected Behavior: - First time load acts normal. Observe the following for few seconds: ![image](https://user-images.githubusercontent.com/29249348/69856559-31e57a00-128e-11ea-9df4-413e3820fc51.png) After a while it fixes itself as following: ![image](https://user-images.githubusercontent.com/29249348/69856652-6822f980-128e-11ea-931b-9bafe90790e4.png)
1.0
Garbled UI Elements on Chrome/Windows on first load. - **Repro steps** - either ctrl-F5, or do a first time load of home page. - Issue resolves after first slider action. Expected Behavior: - First time load acts normal. Observe the following for few seconds: ![image](https://user-images.githubusercontent.com/29249348/69856559-31e57a00-128e-11ea-9df4-413e3820fc51.png) After a while it fixes itself as following: ![image](https://user-images.githubusercontent.com/29249348/69856652-6822f980-128e-11ea-931b-9bafe90790e4.png)
priority
garbled ui elements on chrome windows on first load repro steps either ctrl or do a first time load of home page issue resolves after first slider action expected behavior first time load acts normal observe the following for few seconds after a while it fixes itself as following
1
257,423
8,137,199,727
IssuesEvent
2018-08-20 10:55:21
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.ebay.com - see bug description
browser-firefox-tablet priority-critical
<!-- @browser: Firefox Mobile (Tablet) 63.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0.1; Tablet; rv:63.0) Gecko/63.0 Firefox/63.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.ebay.com/itm/Bayer-Seresto-Flea-and-Tick-Collar-for-cat-8-Month-Protection-odorless/302831012146?hash=item4682228932:g:TO0AAOSwXn5bdP3X **Browser / Version**: Firefox Mobile (Tablet) 63.0 **Operating System**: Android 6.0.1 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: While on ebay pages load until I click on a thumbnail to view an item. When I do so the page gets an error message"The page isn't redirecting properlyAn error occurred during a connection to www.ebay.com. This problem can sometimes be caused by disabling or refusing to accept cookies. Try Again" while I'm in mobile when I switch to desktop site the pages will properly load. **Steps to Reproduce**: None. But the problem was fixed by using the desktop site _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.ebay.com - see bug description - <!-- @browser: Firefox Mobile (Tablet) 63.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0.1; Tablet; rv:63.0) Gecko/63.0 Firefox/63.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.ebay.com/itm/Bayer-Seresto-Flea-and-Tick-Collar-for-cat-8-Month-Protection-odorless/302831012146?hash=item4682228932:g:TO0AAOSwXn5bdP3X **Browser / Version**: Firefox Mobile (Tablet) 63.0 **Operating System**: Android 6.0.1 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: While on ebay pages load until I click on a thumbnail to view an item. When I do so the page gets an error message"The page isn't redirecting properlyAn error occurred during a connection to www.ebay.com. This problem can sometimes be caused by disabling or refusing to accept cookies. Try Again" while I'm in mobile when I switch to desktop site the pages will properly load. **Steps to Reproduce**: None. But the problem was fixed by using the desktop site _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
see bug description url browser version firefox mobile tablet operating system android tested another browser yes problem type something else description while on ebay pages load until i click on a thumbnail to view an item when i do so the page gets an error message the page isn t redirecting properlyan error occurred during a connection to this problem can sometimes be caused by disabling or refusing to accept cookies try again while i m in mobile when i switch to desktop site the pages will properly load steps to reproduce none but the problem was fixed by using the desktop site from with ❤️
1
73,238
19,602,695,559
IssuesEvent
2022-01-06 04:27:12
carla-simulator/carla
https://api.github.com/repos/carla-simulator/carla
closed
Carla 0.9.9 Docker Image Build Error: Error 403 Forbidden
build system
When I'm trying to build the docker image for carla branch 0.9.9, after running this command `docker build -t carla -f Carla.Dockerfile . --build-arg GIT_BRANCH=0.9.9` The error below will show: ![image](https://user-images.githubusercontent.com/93551582/141040879-534b86ed-be01-4ca1-a13a-cfcb041e5e4e.png) This is caused because Autodesk moved the > fbx202001_fbxsdk_linux resource file to a new address so it is no longer reachable from the original wget command. I am on Ubuntu 18.04. Just in case anyone encountered the same issue, I posted a solution here to resolve this issue: My fix was to change line 20 in MyBuildUtilsDocker.sh in the Docker container to: `wget --user-agent="Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0" -c "${FBXSDK_URL}" -P "${CARLA_DOCKER_UTILS_FOLDER}"` Here are the steps I followed: 1 - In Util/Docker, create a file called MyBuildUtilsDocker.sh 2 - cd .. 3 - run cat BuildTools/BuildUtilsDocker.sh > Docker/MyBuildUtilsDocker.sh 4 - Change line 20 in MyBuildUtilsDocker.sh to what I put above. 5 - cd Docker 6 - change Carla.Dockerfile to: ``` FROM carla-prerequisites:latest ARG GIT_BRANCH USER ue4 RUN cd /home/ue4 && \ if [ -z ${GIT_BRANCH+x} ]; then git clone --depth 1 https://github.com/carla-simulator/carla.git; \ else git clone --depth 1 --branch $GIT_BRANCH https://github.com/carla-simulator/carla.git; fi && \ cd /home/ue4/carla && \ ./Update.sh && \ make CarlaUE4Editor && \ make PythonAPI WORKDIR /home/ue4/carla COPY MyBuildUtilsDocker.sh . RUN cat MyBuildUtilsDocker.sh > Util/BuildTools/BuildUtilsDocker.sh RUN make build.utils RUN make package RUN rm -r /home/ue4/carla/Dist WORKDIR /home/ue4/carla ``` 7 - Run `docker build -t carla -f Carla.Dockerfile . --build-arg GIT_BRANCH=0.9.9`
1.0
Carla 0.9.9 Docker Image Build Error: Error 403 Forbidden - When I'm trying to build the docker image for carla branch 0.9.9, after running this command `docker build -t carla -f Carla.Dockerfile . --build-arg GIT_BRANCH=0.9.9` The error below will show: ![image](https://user-images.githubusercontent.com/93551582/141040879-534b86ed-be01-4ca1-a13a-cfcb041e5e4e.png) This is caused because Autodesk moved the > fbx202001_fbxsdk_linux resource file to a new address so it is no longer reachable from the original wget command. I am on Ubuntu 18.04. Just in case anyone encountered the same issue, I posted a solution here to resolve this issue: My fix was to change line 20 in MyBuildUtilsDocker.sh in the Docker container to: `wget --user-agent="Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:94.0) Gecko/20100101 Firefox/94.0" -c "${FBXSDK_URL}" -P "${CARLA_DOCKER_UTILS_FOLDER}"` Here are the steps I followed: 1 - In Util/Docker, create a file called MyBuildUtilsDocker.sh 2 - cd .. 3 - run cat BuildTools/BuildUtilsDocker.sh > Docker/MyBuildUtilsDocker.sh 4 - Change line 20 in MyBuildUtilsDocker.sh to what I put above. 5 - cd Docker 6 - change Carla.Dockerfile to: ``` FROM carla-prerequisites:latest ARG GIT_BRANCH USER ue4 RUN cd /home/ue4 && \ if [ -z ${GIT_BRANCH+x} ]; then git clone --depth 1 https://github.com/carla-simulator/carla.git; \ else git clone --depth 1 --branch $GIT_BRANCH https://github.com/carla-simulator/carla.git; fi && \ cd /home/ue4/carla && \ ./Update.sh && \ make CarlaUE4Editor && \ make PythonAPI WORKDIR /home/ue4/carla COPY MyBuildUtilsDocker.sh . RUN cat MyBuildUtilsDocker.sh > Util/BuildTools/BuildUtilsDocker.sh RUN make build.utils RUN make package RUN rm -r /home/ue4/carla/Dist WORKDIR /home/ue4/carla ``` 7 - Run `docker build -t carla -f Carla.Dockerfile . --build-arg GIT_BRANCH=0.9.9`
non_priority
carla docker image build error error forbidden when i m trying to build the docker image for carla branch after running this command docker build t carla f carla dockerfile build arg git branch the error below will show this is caused because autodesk moved the fbxsdk linux resource file to a new address so it is no longer reachable from the original wget command i am on ubuntu just in case anyone encountered the same issue i posted a solution here to resolve this issue my fix was to change line in mybuildutilsdocker sh in the docker container to wget user agent mozilla ubuntu linux rv gecko firefox c fbxsdk url p carla docker utils folder here are the steps i followed in util docker create a file called mybuildutilsdocker sh cd run cat buildtools buildutilsdocker sh docker mybuildutilsdocker sh change line in mybuildutilsdocker sh to what i put above cd docker change carla dockerfile to from carla prerequisites latest arg git branch user run cd home if then git clone depth else git clone depth branch git branch fi cd home carla update sh make make pythonapi workdir home carla copy mybuildutilsdocker sh run cat mybuildutilsdocker sh util buildtools buildutilsdocker sh run make build utils run make package run rm r home carla dist workdir home carla run docker build t carla f carla dockerfile build arg git branch
0
14,914
10,226,723,853
IssuesEvent
2019-08-16 18:37:59
terraform-providers/terraform-provider-aws
https://api.github.com/repos/terraform-providers/terraform-provider-aws
closed
Terraform errors out on plan if defined NACL rule missing
bug service/ec2
Terraform errors out if a defined NACL rule is not present when `terraform plan` is run. This appears to be due to a bug in the `findNetworkAclRule` function in the referenced .go file. I only became aware of this because I had googled the phrase "Expected the Network ACL to have Entries". I suspect that the correct behavior in this case is for the function to return `nil, nil` rather than falling all the way through to the "I give up" error case. I can't suggest the exact code fix, only that the existing function doesn't correctly act on the expected NACL rule missing when other NACL rules are present. ### Terraform Version 0.10.8 ### Affected Resource(s) - aws_network_acl_rule This appears to affect any possible aws_network_acl ### Terraform Configuration Files https://www.dropbox.com/s/d3a772eddq1nzz4/sample.zip.gpg?dl=0 ### Debug Output https://gist.github.com/AVALARA-WESPAYNE/f5c2ad4c6ac9759e7137d42da94e287e ### Expected Behavior Creation of missing NACL rule added to terraform plan ### Actual Behavior Terraform exits, reporting error which includes output from AWS API DescribeNetworkAcls request. ### Steps to Reproduce 1. Some time in the distant past, create VPC. 2. Using other means, such as AWS web console, remove one or more rules from a NACL. 3. `terraform plan` 4. Terraform errors out with output similar to what is in gist link. ### References - https://github.com/terraform-providers/terraform-provider-aws/blob/master/aws/resource_aws_network_acl_rule.go
1.0
Terraform errors out on plan if defined NACL rule missing - Terraform errors out if a defined NACL rule is not present when `terraform plan` is run. This appears to be due to a bug in the `findNetworkAclRule` function in the referenced .go file. I only became aware of this because I had googled the phrase "Expected the Network ACL to have Entries". I suspect that the correct behavior in this case is for the function to return `nil, nil` rather than falling all the way through to the "I give up" error case. I can't suggest the exact code fix, only that the existing function doesn't correctly act on the expected NACL rule missing when other NACL rules are present. ### Terraform Version 0.10.8 ### Affected Resource(s) - aws_network_acl_rule This appears to affect any possible aws_network_acl ### Terraform Configuration Files https://www.dropbox.com/s/d3a772eddq1nzz4/sample.zip.gpg?dl=0 ### Debug Output https://gist.github.com/AVALARA-WESPAYNE/f5c2ad4c6ac9759e7137d42da94e287e ### Expected Behavior Creation of missing NACL rule added to terraform plan ### Actual Behavior Terraform exits, reporting error which includes output from AWS API DescribeNetworkAcls request. ### Steps to Reproduce 1. Some time in the distant past, create VPC. 2. Using other means, such as AWS web console, remove one or more rules from a NACL. 3. `terraform plan` 4. Terraform errors out with output similar to what is in gist link. ### References - https://github.com/terraform-providers/terraform-provider-aws/blob/master/aws/resource_aws_network_acl_rule.go
non_priority
terraform errors out on plan if defined nacl rule missing terraform errors out if a defined nacl rule is not present when terraform plan is run this appears to be due to a bug in the findnetworkaclrule function in the referenced go file i only became aware of this because i had googled the phrase expected the network acl to have entries i suspect that the correct behavior in this case is for the function to return nil nil rather than falling all the way through to the i give up error case i can t suggest the exact code fix only that the existing function doesn t correctly act on the expected nacl rule missing when other nacl rules are present terraform version affected resource s aws network acl rule this appears to affect any possible aws network acl terraform configuration files debug output expected behavior creation of missing nacl rule added to terraform plan actual behavior terraform exits reporting error which includes output from aws api describenetworkacls request steps to reproduce some time in the distant past create vpc using other means such as aws web console remove one or more rules from a nacl terraform plan terraform errors out with output similar to what is in gist link references
0
673,637
23,024,494,276
IssuesEvent
2022-07-22 08:08:57
owncloud/web
https://api.github.com/repos/owncloud/web
closed
Safari 14.1 + 15.5: rightclick-context menu appears always next to 3-dots-quickaction
Type:Bug Priority:p2-high GA-Blocker
### Steps to reproduce 1. login to https://ocis.ocis-web.latest.owncloud.works/ 2. create a resource 3. rightlick on the resource 4. rightclick-context menu appears always next to 3-dots-quickaction https://user-images.githubusercontent.com/26610733/179959640-94dc4dbb-652f-48d9-8ba4-4eeecc3d413f.mp4 ### Expected behaviour context-menu should appear next to the mouse pointer ### Actual behaviour rightclick-context menu appears always next to 3-dots-quickaction ### Environment general - Safari 14.1 (M1 Silicon) - Safari 15.5 (Intel)
1.0
Safari 14.1 + 15.5: rightclick-context menu appears always next to 3-dots-quickaction - ### Steps to reproduce 1. login to https://ocis.ocis-web.latest.owncloud.works/ 2. create a resource 3. rightlick on the resource 4. rightclick-context menu appears always next to 3-dots-quickaction https://user-images.githubusercontent.com/26610733/179959640-94dc4dbb-652f-48d9-8ba4-4eeecc3d413f.mp4 ### Expected behaviour context-menu should appear next to the mouse pointer ### Actual behaviour rightclick-context menu appears always next to 3-dots-quickaction ### Environment general - Safari 14.1 (M1 Silicon) - Safari 15.5 (Intel)
priority
safari rightclick context menu appears always next to dots quickaction steps to reproduce login to create a resource rightlick on the resource rightclick context menu appears always next to dots quickaction expected behaviour context menu should appear next to the mouse pointer actual behaviour rightclick context menu appears always next to dots quickaction environment general safari silicon safari intel
1
193,798
6,888,147,855
IssuesEvent
2017-11-22 03:49:45
HoneycuttInc/Thorncastle
https://api.github.com/repos/HoneycuttInc/Thorncastle
closed
Task 3.1: Just closed IE only to reopen it now
Low Priority
Last task ended with closing IE, this one begins with opening it. If this is necessary, maybe explain why?
1.0
Task 3.1: Just closed IE only to reopen it now - Last task ended with closing IE, this one begins with opening it. If this is necessary, maybe explain why?
priority
task just closed ie only to reopen it now last task ended with closing ie this one begins with opening it if this is necessary maybe explain why
1
4,785
7,661,273,284
IssuesEvent
2018-05-11 13:44:51
SharePoint/PnP-PowerShell
https://api.github.com/repos/SharePoint/PnP-PowerShell
closed
Add-PnPFile can't write Values Modified & Created
Needs investigation To be processed
### Reporting an Issue or Missing Feature When uploading a file with Add-PnPFile values for Modified & Created are ignored ### Expected behavior > PS > $pubdate > **Thursday, 19 November 2015 2:47:38 PM** > > PS > $pubdate.GetType() > IsPublic IsSerial Name BaseType > True True **DateTime** System.ValueType > > PS > Add-PnPFile -Path $path -Folder $folder -Values @{Modified="$pubdate"; Created="$pubdate"} > Name Type Items/Size Last Modified > 2015_07 Annual Leave.pdf File 57000 **19/11/2015 2:47:38 PM** ### Actual behavior > PS > $pubdate > **Thursday, 19 November 2015 2:47:38 PM** > > PS > $pubdate.GetType() > IsPublic IsSerial Name BaseType > True True **DateTime** System.ValueType > > PS > Add-PnPFile -Path $path -Folder $folder -Values @{Modified="$pubdate"; Created="$pubdate"} > Name Type Items/Size Last Modified > 2015_07 Annual Leave.pdf File 57000 **7/05/2018 5:49:01 AM** ### Steps to reproduce behavior Happens with every file upload for me, I can't set the modified or created date no matter what I do. However I can create a new Date & Time column and set that just fine. ### Which version of the PnP-PowerShell Cmdlets are you using? - [ ] PnP PowerShell for SharePoint 2013 - [ ] PnP PowerShell for SharePoint 2016 - [x] PnP PowerShell for SharePoint Online ### What is the version of the Cmdlet module you are running? SharePointPnPPowerShellOnline 2.25.1804.1 ### How did you install the PnP-PowerShell Cmdlets? - [ ] MSI Installed downloaded from GitHub - [x] Installed through the PowerShell Gallery with Install-Module - [ ] Other means
1.0
Add-PnPFile can't write Values Modified & Created - ### Reporting an Issue or Missing Feature When uploading a file with Add-PnPFile values for Modified & Created are ignored ### Expected behavior > PS > $pubdate > **Thursday, 19 November 2015 2:47:38 PM** > > PS > $pubdate.GetType() > IsPublic IsSerial Name BaseType > True True **DateTime** System.ValueType > > PS > Add-PnPFile -Path $path -Folder $folder -Values @{Modified="$pubdate"; Created="$pubdate"} > Name Type Items/Size Last Modified > 2015_07 Annual Leave.pdf File 57000 **19/11/2015 2:47:38 PM** ### Actual behavior > PS > $pubdate > **Thursday, 19 November 2015 2:47:38 PM** > > PS > $pubdate.GetType() > IsPublic IsSerial Name BaseType > True True **DateTime** System.ValueType > > PS > Add-PnPFile -Path $path -Folder $folder -Values @{Modified="$pubdate"; Created="$pubdate"} > Name Type Items/Size Last Modified > 2015_07 Annual Leave.pdf File 57000 **7/05/2018 5:49:01 AM** ### Steps to reproduce behavior Happens with every file upload for me, I can't set the modified or created date no matter what I do. However I can create a new Date & Time column and set that just fine. ### Which version of the PnP-PowerShell Cmdlets are you using? - [ ] PnP PowerShell for SharePoint 2013 - [ ] PnP PowerShell for SharePoint 2016 - [x] PnP PowerShell for SharePoint Online ### What is the version of the Cmdlet module you are running? SharePointPnPPowerShellOnline 2.25.1804.1 ### How did you install the PnP-PowerShell Cmdlets? - [ ] MSI Installed downloaded from GitHub - [x] Installed through the PowerShell Gallery with Install-Module - [ ] Other means
non_priority
add pnpfile can t write values modified created reporting an issue or missing feature when uploading a file with add pnpfile values for modified created are ignored expected behavior ps pubdate thursday november pm ps pubdate gettype ispublic isserial name basetype true true datetime system valuetype ps add pnpfile path path folder folder values modified pubdate created pubdate name type items size last modified annual leave pdf file pm actual behavior ps pubdate thursday november pm ps pubdate gettype ispublic isserial name basetype true true datetime system valuetype ps add pnpfile path path folder folder values modified pubdate created pubdate name type items size last modified annual leave pdf file am steps to reproduce behavior happens with every file upload for me i can t set the modified or created date no matter what i do however i can create a new date time column and set that just fine which version of the pnp powershell cmdlets are you using pnp powershell for sharepoint pnp powershell for sharepoint pnp powershell for sharepoint online what is the version of the cmdlet module you are running sharepointpnppowershellonline how did you install the pnp powershell cmdlets msi installed downloaded from github installed through the powershell gallery with install module other means
0
192,134
15,339,730,145
IssuesEvent
2021-02-27 03:17:39
UBC-MDS/picturepyfect
https://api.github.com/repos/UBC-MDS/picturepyfect
closed
README.md
documentation
Outline the package you would like to build in the README.md file. (This can be identical for both projects at this point in the project). In particular, your README.md should contain: - a summary paragraph that describes the project at a high level - a bulleted list of the functions (and datasets if applicable) that will be included in the package (this should be a 1-2 sentence description for each function/dataset) - a paragraph describing where your packages fit into the Python ecosystem (are there any other Python packages that have the same/similar functionality? Provide links to any that do. If none exist, then clearly state this as well).
1.0
README.md - Outline the package you would like to build in the README.md file. (This can be identical for both projects at this point in the project). In particular, your README.md should contain: - a summary paragraph that describes the project at a high level - a bulleted list of the functions (and datasets if applicable) that will be included in the package (this should be a 1-2 sentence description for each function/dataset) - a paragraph describing where your packages fit into the Python ecosystem (are there any other Python packages that have the same/similar functionality? Provide links to any that do. If none exist, then clearly state this as well).
non_priority
readme md outline the package you would like to build in the readme md file this can be identical for both projects at this point in the project in particular your readme md should contain a summary paragraph that describes the project at a high level a bulleted list of the functions and datasets if applicable that will be included in the package this should be a sentence description for each function dataset a paragraph describing where your packages fit into the python ecosystem are there any other python packages that have the same similar functionality provide links to any that do if none exist then clearly state this as well
0
453,398
13,069,264,924
IssuesEvent
2020-07-31 06:05:53
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.youtube.com - site is not usable
browser-firefox-mobile engine-gecko priority-critical
<!-- @browser: Firefox Mobile 80.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:80.0) Gecko/80.0 Firefox/80.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/56046 --> **URL**: https://www.youtube.com/ **Browser / Version**: Firefox Mobile 80.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: When I open YouTube site, the browser need me to open it in another browser <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200724093206</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> Submitted in the name of `@allo0osh` _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.youtube.com - site is not usable - <!-- @browser: Firefox Mobile 80.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:80.0) Gecko/80.0 Firefox/80.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/56046 --> **URL**: https://www.youtube.com/ **Browser / Version**: Firefox Mobile 80.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: When I open YouTube site, the browser need me to open it in another browser <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200724093206</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> Submitted in the name of `@allo0osh` _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description browser unsupported steps to reproduce when i open youtube site the browser need me to open it in another browser browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true submitted in the name of from with ❤️
1
244,065
7,870,331,392
IssuesEvent
2018-06-25 00:32:39
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Unable to get deployment logs for multi container pods
component/cli kind/bug lifecycle/rotten priority/P2
When there are multiple containers in pod `oc logs` fails for deploymentconfigs for that pod and it is not possible to specify the container with -c parameter. ##### Version oc v3.2.0.20 kubernetes v1.2.0-36-g4a3f9c5 ##### Steps To Reproduce 1. oc logs dc/name_of_pod_with_multiple_containers -c name_of_container ##### Current Result Error from server: a container name must be specified for pod ..., choose one of: [... ...] ##### Expected Result logs...
1.0
Unable to get deployment logs for multi container pods - When there are multiple containers in pod `oc logs` fails for deploymentconfigs for that pod and it is not possible to specify the container with -c parameter. ##### Version oc v3.2.0.20 kubernetes v1.2.0-36-g4a3f9c5 ##### Steps To Reproduce 1. oc logs dc/name_of_pod_with_multiple_containers -c name_of_container ##### Current Result Error from server: a container name must be specified for pod ..., choose one of: [... ...] ##### Expected Result logs...
priority
unable to get deployment logs for multi container pods when there are multiple containers in pod oc logs fails for deploymentconfigs for that pod and it is not possible to specify the container with c parameter version oc kubernetes steps to reproduce oc logs dc name of pod with multiple containers c name of container current result error from server a container name must be specified for pod choose one of expected result logs
1
21,982
6,227,830,899
IssuesEvent
2017-07-10 21:38:47
XceedBoucherS/TestImport5
https://api.github.com/repos/XceedBoucherS/TestImport5
closed
Problem with BusyIndicator
CodePlex
<b>janglinj[CodePlex]</b> <br />I'm having a problem where after the busy indicator is done it leaves some of my buttons with disabled appearance. They look disabled but they still work. Any chance of fixing that? See the attachment. The image on the left is before the busyindicator was displayed. The image is after the busyindicator was displayed and hidden by setting IsBusy to true and then false.
1.0
Problem with BusyIndicator - <b>janglinj[CodePlex]</b> <br />I'm having a problem where after the busy indicator is done it leaves some of my buttons with disabled appearance. They look disabled but they still work. Any chance of fixing that? See the attachment. The image on the left is before the busyindicator was displayed. The image is after the busyindicator was displayed and hidden by setting IsBusy to true and then false.
non_priority
problem with busyindicator janglinj i m having a problem where after the busy indicator is done it leaves some of my buttons with disabled appearance they look disabled but they still work any chance of fixing that see the attachment the image on the left is before the busyindicator was displayed the image is after the busyindicator was displayed and hidden by setting isbusy to true and then false
0
459,910
13,201,119,938
IssuesEvent
2020-08-14 09:31:20
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
uart_fifo_read can only read one character
bug platform: STM32 priority: low
**Describe the bug** ''' int uart_fifo_init(void){ uint8_t c; dev = device_get_binding("UART_2"); if (dev == NULL) { printk("cannot find device\n"); return -1; } uart_irq_rx_disable(dev); uart_irq_tx_disable(dev); uart_irq_callback_set(dev, uart_fifo_callback); /* Drain the fifo */ if (uart_irq_rx_ready(dev)) { while (uart_fifo_read(dev, &c, 1)) { ; } } uart_irq_rx_enable(dev); /* Enable all interrupts unconditionally. Note that this is due * to Zephyr issue #8393. This should be removed once the * issue is fixed in upstream Zephyr. */ irq_unlock(0); return 0; } void uart_fifo_callback(struct device *dev){ while(uart_irq_update(dev) && uart_irq_is_pending(dev)){ if (!uart_irq_rx_ready(dev)){ continue; } else{ buf.size += uart_fifo_read(dev, &buf.buf[buf.size], 10); printk("buf.size:%d\n",buf.size); } printk("while1\n"); } } ''' when I set up the callback function as above and input a line of characters by serial to uart2 rx, it can only read the first character **Expected behavior** A clear and concise description of what you expected to happen. If I input "aaa" to uart2, my buf should contain 0x61,0x61,0x61. But now it only contains one 0x61. **Logs and console output** If applicable, add console logs or other types of debug information e.g Wireshark capture or Logic analyzer capture (upload in zip archive). copy-and-paste text and put a code fence (\`\`\`) before and after, to help explain the issue. (if unable to obtain text log, add a screenshot) ![Snip20200806_5](https://user-images.githubusercontent.com/37931674/89519781-76f38000-d80f-11ea-8fd3-dca89b085df0.png) **Environment (please complete the following information):** - OS: MacOS - Toolchain: gcc-arm-none-eabi
1.0
uart_fifo_read can only read one character - **Describe the bug** ''' int uart_fifo_init(void){ uint8_t c; dev = device_get_binding("UART_2"); if (dev == NULL) { printk("cannot find device\n"); return -1; } uart_irq_rx_disable(dev); uart_irq_tx_disable(dev); uart_irq_callback_set(dev, uart_fifo_callback); /* Drain the fifo */ if (uart_irq_rx_ready(dev)) { while (uart_fifo_read(dev, &c, 1)) { ; } } uart_irq_rx_enable(dev); /* Enable all interrupts unconditionally. Note that this is due * to Zephyr issue #8393. This should be removed once the * issue is fixed in upstream Zephyr. */ irq_unlock(0); return 0; } void uart_fifo_callback(struct device *dev){ while(uart_irq_update(dev) && uart_irq_is_pending(dev)){ if (!uart_irq_rx_ready(dev)){ continue; } else{ buf.size += uart_fifo_read(dev, &buf.buf[buf.size], 10); printk("buf.size:%d\n",buf.size); } printk("while1\n"); } } ''' when I set up the callback function as above and input a line of characters by serial to uart2 rx, it can only read the first character **Expected behavior** A clear and concise description of what you expected to happen. If I input "aaa" to uart2, my buf should contain 0x61,0x61,0x61. But now it only contains one 0x61. **Logs and console output** If applicable, add console logs or other types of debug information e.g Wireshark capture or Logic analyzer capture (upload in zip archive). copy-and-paste text and put a code fence (\`\`\`) before and after, to help explain the issue. (if unable to obtain text log, add a screenshot) ![Snip20200806_5](https://user-images.githubusercontent.com/37931674/89519781-76f38000-d80f-11ea-8fd3-dca89b085df0.png) **Environment (please complete the following information):** - OS: MacOS - Toolchain: gcc-arm-none-eabi
priority
uart fifo read can only read one character describe the bug int uart fifo init void t c dev device get binding uart if dev null printk cannot find device n return uart irq rx disable dev uart irq tx disable dev uart irq callback set dev uart fifo callback drain the fifo if uart irq rx ready dev while uart fifo read dev c uart irq rx enable dev enable all interrupts unconditionally note that this is due to zephyr issue this should be removed once the issue is fixed in upstream zephyr irq unlock return void uart fifo callback struct device dev while uart irq update dev uart irq is pending dev if uart irq rx ready dev continue else buf size uart fifo read dev buf buf printk buf size d n buf size printk n when i set up the callback function as above and input a line of characters by serial to rx it can only read the first character expected behavior a clear and concise description of what you expected to happen if i input aaa to my buf should contain but now it only contains one logs and console output if applicable add console logs or other types of debug information e g wireshark capture or logic analyzer capture upload in zip archive copy and paste text and put a code fence before and after to help explain the issue if unable to obtain text log add a screenshot environment please complete the following information os macos toolchain gcc arm none eabi
1
657,693
21,801,282,557
IssuesEvent
2022-05-16 05:42:03
Automattic/woocommerce-payments
https://api.github.com/repos/Automattic/woocommerce-payments
closed
Display fee breakdown in order notes.
type: enhancement priority: high status: has pr component: order details size: medium component: countries currencies localization impact: high category: core
Splitted from: https://github.com/Automattic/woocommerce-payments/issues/2233 All transactions are subject to a base fee (determined by the merchant's country) plus a number of additional fees (like international card processing, FX fees etc). We currently only show the total fee. We want to break this number down to make it clearer how the final rate was calculated. **Order Notes** ![image](https://user-images.githubusercontent.com/686419/122224470-0c3db700-ceac-11eb-8b4b-eedaf607a406.png)
1.0
Display fee breakdown in order notes. - Splitted from: https://github.com/Automattic/woocommerce-payments/issues/2233 All transactions are subject to a base fee (determined by the merchant's country) plus a number of additional fees (like international card processing, FX fees etc). We currently only show the total fee. We want to break this number down to make it clearer how the final rate was calculated. **Order Notes** ![image](https://user-images.githubusercontent.com/686419/122224470-0c3db700-ceac-11eb-8b4b-eedaf607a406.png)
priority
display fee breakdown in order notes splitted from all transactions are subject to a base fee determined by the merchant s country plus a number of additional fees like international card processing fx fees etc we currently only show the total fee we want to break this number down to make it clearer how the final rate was calculated order notes
1
284,658
21,464,216,290
IssuesEvent
2022-04-26 00:46:25
makeworld-the-better-one/amfora
https://api.github.com/repos/makeworld-the-better-one/amfora
closed
flatpak package for Amfora
documentation enhancement help wanted
Would really like to have a flatpak package for Amfora (or a .deb package but I think flatpak would cover more ground).
1.0
flatpak package for Amfora - Would really like to have a flatpak package for Amfora (or a .deb package but I think flatpak would cover more ground).
non_priority
flatpak package for amfora would really like to have a flatpak package for amfora or a deb package but i think flatpak would cover more ground
0
2,209
24,150,830,507
IssuesEvent
2022-09-22 00:24:57
hackforla/ops
https://api.github.com/repos/hackforla/ops
opened
Create Glossary of AWS resources used in Incubator
size: 1pt role: Site Reliability Engineer feature: administrative
### Overview ![image](https://user-images.githubusercontent.com/6300995/191632501-dd31ae05-b996-4bdf-99dd-2496de333df3.png) As a Dev Ops member, I would like to have a glossary or list of all active AWS resources used by Incubator to deploy its apps. These resources can be identified within the Incubator's code in the [repo](https://github.com/hackforla/incubator/tree/main/terraform-modules). _CONFIRMATION_ As a github user, when I click on [Incubator's wiki page](https://github.com/hackforla/incubator/wiki), I should see an article AWS Resources Inside should be a document of a list of aws resources (Route 53, Amazon ECR, etc.) with links or references to their AWS glossary. Let's also consider linking to the Incubator's terraform service code as well. ### Action Items - [ ] [Examine Incubator's infrastructure code](https://github.com/hackforla/incubator/tree/main/terraform-modules) - [ ] Create AWS resources on Incubator wiki ### Resources/Instructions [Incubator repo](https://github.com/hackforla/incubator/tree/main/terraform-modules) [Wiki](https://github.com/hackforla/incubator/wiki)
True
Create Glossary of AWS resources used in Incubator - ### Overview ![image](https://user-images.githubusercontent.com/6300995/191632501-dd31ae05-b996-4bdf-99dd-2496de333df3.png) As a Dev Ops member, I would like to have a glossary or list of all active AWS resources used by Incubator to deploy its apps. These resources can be identified within the Incubator's code in the [repo](https://github.com/hackforla/incubator/tree/main/terraform-modules). _CONFIRMATION_ As a github user, when I click on [Incubator's wiki page](https://github.com/hackforla/incubator/wiki), I should see an article AWS Resources Inside should be a document of a list of aws resources (Route 53, Amazon ECR, etc.) with links or references to their AWS glossary. Let's also consider linking to the Incubator's terraform service code as well. ### Action Items - [ ] [Examine Incubator's infrastructure code](https://github.com/hackforla/incubator/tree/main/terraform-modules) - [ ] Create AWS resources on Incubator wiki ### Resources/Instructions [Incubator repo](https://github.com/hackforla/incubator/tree/main/terraform-modules) [Wiki](https://github.com/hackforla/incubator/wiki)
non_priority
create glossary of aws resources used in incubator overview as a dev ops member i would like to have a glossary or list of all active aws resources used by incubator to deploy its apps these resources can be identified within the incubator s code in the confirmation as a github user when i click on i should see an article aws resources inside should be a document of a list of aws resources route amazon ecr etc with links or references to their aws glossary let s also consider linking to the incubator s terraform service code as well action items create aws resources on incubator wiki resources instructions
0
692,393
23,732,745,443
IssuesEvent
2022-08-31 04:23:05
wso2/api-manager
https://api.github.com/repos/wso2/api-manager
opened
Make MGW default cache expiry time as a configurable property from MGW level
Type/Bug Priority/Normal
### Description Currently, in MGW, when a token comes with a non integer "exp" value, the Oauth2 token cache is set to a default expiry time with 1 hour. What the customer expect is, since we can configure cache expiry time in the micro-gw.conf file, regardless of the value of "exp" value of the token, it should be set to the configured value, not to any default. So as an alternative approach, we can make this default cache expiry time as a configurable property from Micro GW level. ### Steps to Reproduce - Configured APIM with MGW. - Created an API in APIM 3.2. - Cache expiry has been set to 15min - Imported this API to the MGW toolkit and added the jar to the MGW runtime. - Then I tried to call this with an opaque token ( The expiry time is coming as a string not as an Integer) - The cache was not expired in 15 min, instead it was 1 hour. ### Affected Component APIM ### Version 3.2.0 ### Environment Details (with versions) _No response_ ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels MGW
1.0
Make MGW default cache expiry time as a configurable property from MGW level - ### Description Currently, in MGW, when a token comes with a non integer "exp" value, the Oauth2 token cache is set to a default expiry time with 1 hour. What the customer expect is, since we can configure cache expiry time in the micro-gw.conf file, regardless of the value of "exp" value of the token, it should be set to the configured value, not to any default. So as an alternative approach, we can make this default cache expiry time as a configurable property from Micro GW level. ### Steps to Reproduce - Configured APIM with MGW. - Created an API in APIM 3.2. - Cache expiry has been set to 15min - Imported this API to the MGW toolkit and added the jar to the MGW runtime. - Then I tried to call this with an opaque token ( The expiry time is coming as a string not as an Integer) - The cache was not expired in 15 min, instead it was 1 hour. ### Affected Component APIM ### Version 3.2.0 ### Environment Details (with versions) _No response_ ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels MGW
priority
make mgw default cache expiry time as a configurable property from mgw level description currently in mgw when a token comes with a non integer exp value the token cache is set to a default expiry time with hour what the customer expect is since we can configure cache expiry time in the micro gw conf file regardless of the value of exp value of the token it should be set to the configured value not to any default so as an alternative approach we can make this default cache expiry time as a configurable property from micro gw level steps to reproduce configured apim with mgw created an api in apim cache expiry has been set to imported this api to the mgw toolkit and added the jar to the mgw runtime then i tried to call this with an opaque token the expiry time is coming as a string not as an integer the cache was not expired in min instead it was hour affected component apim version environment details with versions no response relevant log output no response related issues no response suggested labels mgw
1
29,472
2,716,125,483
IssuesEvent
2015-04-10 17:11:00
CruxFramework/crux
https://api.github.com/repos/CruxFramework/crux
closed
Hide or Show Slider controls
bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.1.1
_From [samuel@cruxframework.org](https://code.google.com/u/samuel@cruxframework.org/) on June 11, 2014 15:05:39_ Add a method to implement this behavior _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=398_
1.0
Hide or Show Slider controls - _From [samuel@cruxframework.org](https://code.google.com/u/samuel@cruxframework.org/) on June 11, 2014 15:05:39_ Add a method to implement this behavior _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=398_
priority
hide or show slider controls from on june add a method to implement this behavior original issue
1
132,772
12,517,553,229
IssuesEvent
2020-06-03 11:22:11
abrt/abrt
https://api.github.com/repos/abrt/abrt
closed
update links of fedorahosted.org/abrt
documentation
Most of these files need to be updated. https://github.com/search?utf8=%E2%9C%93&q=org%3Aabrt+%22fedorahosted.org%2Fabrt%22&type=Code And we probably can update strings in translations. I.e. download updates from zanata, change url, upload translation and sources to zanata.
1.0
update links of fedorahosted.org/abrt - Most of these files need to be updated. https://github.com/search?utf8=%E2%9C%93&q=org%3Aabrt+%22fedorahosted.org%2Fabrt%22&type=Code And we probably can update strings in translations. I.e. download updates from zanata, change url, upload translation and sources to zanata.
non_priority
update links of fedorahosted org abrt most of these files need to be updated and we probably can update strings in translations i e download updates from zanata change url upload translation and sources to zanata
0
252,900
27,271,379,909
IssuesEvent
2023-02-22 22:43:13
snowflakedb/libsnowflakeclient
https://api.github.com/repos/snowflakedb/libsnowflakeclient
closed
util-linuxv2.36.2: 2 vulnerabilities (highest severity is: 5.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>util-linuxv2.36.2</b></p></summary> <p> <p>The util-linux code repository.</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p> </p> </p></p> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b> </p> <p></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-0563](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | util-linuxv2.36.2 | Direct | v2.37.4 | &#10060; | | [CVE-2021-3995](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.7 | util-linuxv2.36.2 | Direct | v2.37.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0563</summary> ### Vulnerable Library - <b>util-linuxv2.36.2</b></p> <p> <p>The util-linux code repository.</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p> <p>Found in base branch: <b>master</b></p></p> </p></p> ### Vulnerable Source Files (2) <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b> </p> <p></p> </p> <p></p> ### Vulnerability Details <p> A flaw was found in the util-linux chfn and chsh utilities when compiled with Readline support. The Readline library uses an "INPUTRC" environment variable to get a path to the library config file. When the library cannot parse the specified file, it prints an error message containing data from the file. This flaw allows an unprivileged user to read root-owned files, potentially leading to privilege escalation. This flaw affects util-linux versions prior to 2.37.4. <p>Publish Date: 2022-02-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563>CVE-2022-0563</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=2053151">https://bugzilla.redhat.com/show_bug.cgi?id=2053151</a></p> <p>Release Date: 2022-02-21</p> <p>Fix Resolution: v2.37.4</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-3995</summary> ### Vulnerable Library - <b>util-linuxv2.36.2</b></p> <p> <p>The util-linux code repository.</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p> <p>Found in base branch: <b>master</b></p></p> </p></p> ### Vulnerable Source Files (1) <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/libmount/src/mountP.h</b> </p> <p></p> </p> <p></p> ### Vulnerability Details <p> A logic error was found in the libmount library of util-linux in the function that allows an unprivileged user to unmount a FUSE filesystem. This flaw allows an unprivileged local attacker to unmount FUSE filesystems that belong to certain other users who have a UID that is a prefix of the UID of the attacker in its string form. An attacker may use this flaw to cause a denial of service to applications that use the affected filesystems. <p>Publish Date: 2021-11-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995>CVE-2021-3995</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>4.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2021-3995">https://security-tracker.debian.org/tracker/CVE-2021-3995</a></p> <p>Release Date: 2021-11-22</p> <p>Fix Resolution: v2.37.3</p> </p> <p></p> </details>
True
util-linuxv2.36.2: 2 vulnerabilities (highest severity is: 5.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>util-linuxv2.36.2</b></p></summary> <p> <p>The util-linux code repository.</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p> </p> </p></p> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b> </p> <p></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-0563](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | util-linuxv2.36.2 | Direct | v2.37.4 | &#10060; | | [CVE-2021-3995](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.7 | util-linuxv2.36.2 | Direct | v2.37.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0563</summary> ### Vulnerable Library - <b>util-linuxv2.36.2</b></p> <p> <p>The util-linux code repository.</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p> <p>Found in base branch: <b>master</b></p></p> </p></p> ### Vulnerable Source Files (2) <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b> </p> <p></p> </p> <p></p> ### Vulnerability Details <p> A flaw was found in the util-linux chfn and chsh utilities when compiled with Readline support. The Readline library uses an "INPUTRC" environment variable to get a path to the library config file. When the library cannot parse the specified file, it prints an error message containing data from the file. This flaw allows an unprivileged user to read root-owned files, potentially leading to privilege escalation. This flaw affects util-linux versions prior to 2.37.4. <p>Publish Date: 2022-02-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563>CVE-2022-0563</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=2053151">https://bugzilla.redhat.com/show_bug.cgi?id=2053151</a></p> <p>Release Date: 2022-02-21</p> <p>Fix Resolution: v2.37.4</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-3995</summary> ### Vulnerable Library - <b>util-linuxv2.36.2</b></p> <p> <p>The util-linux code repository.</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p> <p>Found in base branch: <b>master</b></p></p> </p></p> ### Vulnerable Source Files (1) <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/libmount/src/mountP.h</b> </p> <p></p> </p> <p></p> ### Vulnerability Details <p> A logic error was found in the libmount library of util-linux in the function that allows an unprivileged user to unmount a FUSE filesystem. This flaw allows an unprivileged local attacker to unmount FUSE filesystems that belong to certain other users who have a UID that is a prefix of the UID of the attacker in its string form. An attacker may use this flaw to cause a denial of service to applications that use the affected filesystems. <p>Publish Date: 2021-11-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995>CVE-2021-3995</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>4.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2021-3995">https://security-tracker.debian.org/tracker/CVE-2021-3995</a></p> <p>Release Date: 2021-11-22</p> <p>Fix Resolution: v2.37.3</p> </p> <p></p> </details>
non_priority
util vulnerabilities highest severity is vulnerable library util the util linux code repository library home page a href vulnerable source files deps util linux tar util linux login utils chsh c deps util linux tar util linux login utils chfn c vulnerabilities cve severity cvss dependency type fixed in remediation available medium util direct medium util direct details cve vulnerable library util the util linux code repository library home page a href found in base branch master vulnerable source files deps util linux tar util linux login utils chsh c deps util linux tar util linux login utils chfn c vulnerability details a flaw was found in the util linux chfn and chsh utilities when compiled with readline support the readline library uses an inputrc environment variable to get a path to the library config file when the library cannot parse the specified file it prints an error message containing data from the file this flaw allows an unprivileged user to read root owned files potentially leading to privilege escalation this flaw affects util linux versions prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cve vulnerable library util the util linux code repository library home page a href found in base branch master vulnerable source files deps util linux tar util linux libmount src mountp h vulnerability details a logic error was found in the libmount library of util linux in the function that allows an unprivileged user to unmount a fuse filesystem this flaw allows an unprivileged local attacker to unmount fuse filesystems that belong to certain other users who have a uid that is a prefix of the uid of the attacker in its string form an attacker may use this flaw to cause a denial of service to applications that use the affected filesystems publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
702,693
24,131,973,904
IssuesEvent
2022-09-21 08:13:12
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
barclays.co.uk - site is not usable
browser-firefox priority-normal engine-gecko
<!-- @browser: Firefox 104.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:104.0) Gecko/20100101 Firefox/104.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/111122 --> **URL**: https://barclays.co.uk **Browser / Version**: Firefox 104.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Unable to login **Steps to Reproduce**: Log in first page (using card number log in and name) works OK. Next log in page using pin-sentry does not work. You are required to enter last four digits of card number but the field into which this number is entered does not appear. It seems to be covered by a distorted graphic of the pin sentry machine. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
barclays.co.uk - site is not usable - <!-- @browser: Firefox 104.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:104.0) Gecko/20100101 Firefox/104.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/111122 --> **URL**: https://barclays.co.uk **Browser / Version**: Firefox 104.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Unable to login **Steps to Reproduce**: Log in first page (using card number log in and name) works OK. Next log in page using pin-sentry does not work. You are required to enter last four digits of card number but the field into which this number is entered does not appear. It seems to be covered by a distorted graphic of the pin sentry machine. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
barclays co uk site is not usable url browser version firefox operating system windows tested another browser yes edge problem type site is not usable description unable to login steps to reproduce log in first page using card number log in and name works ok next log in page using pin sentry does not work you are required to enter last four digits of card number but the field into which this number is entered does not appear it seems to be covered by a distorted graphic of the pin sentry machine browser configuration none from with ❤️
1
180,224
6,647,422,911
IssuesEvent
2017-09-28 03:53:28
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio-ui] Some items in various dialogs are not translated when the language is changed
enhancement Priority: Medium
Some items in various dialogs are not translated when the language is changed. Using the website_editorial bp, change the language by clicking on your user name at the top right of Studio, then select **Settings** Under **Profile**, change the language from **English** to **Spanish**, then click on the **Update Profile** button. Here are the dialogs containing items that may need translation: **Account Management** ![screen shot 2017-09-20 at 10 42 44 am](https://user-images.githubusercontent.com/25483966/30650411-cc795204-9df0-11e7-8a22-5aede16ef674.png) **My Sites** ![screen shot 2017-09-20 at 10 45 40 am](https://user-images.githubusercontent.com/25483966/30650477-fb34e0f4-9df0-11e7-8d80-f79e007c3a14.png) **Dashboard Icon Guide** ![screen shot 2017-09-20 at 10 48 23 am](https://user-images.githubusercontent.com/25483966/30650729-af07d2b2-9df1-11e7-8283-a50ce072e913.png) **About** ![screen shot 2017-09-20 at 10 52 32 am](https://user-images.githubusercontent.com/25483966/30650838-e9eda122-9df1-11e7-91f8-b641a4f77a76.png) **Users** ![screen shot 2017-09-20 at 12 09 45 pm](https://user-images.githubusercontent.com/25483966/30654735-c6488f24-9dfc-11e7-940a-7151612ae53e.png) **Create User** ![screen shot 2017-09-20 at 12 07 37 pm](https://user-images.githubusercontent.com/25483966/30654635-7a2b73d6-9dfc-11e7-90fb-1a0d2fac7827.png) **Publish Status** <img width="373" alt="screen shot 2017-09-20 at 4 12 16 pm" src="https://user-images.githubusercontent.com/25483966/30665374-c798e2a8-9e1e-11e7-82da-0531197f98be.png"> **Targeting** <img width="728" alt="screen shot 2017-09-20 at 4 15 29 pm" src="https://user-images.githubusercontent.com/25483966/30665500-2c4741cc-9e1f-11e7-8cdb-de7ed9b92a03.png"> **Edit Form** <img width="1131" alt="screen shot 2017-09-20 at 4 19 44 pm" src="https://user-images.githubusercontent.com/25483966/30665792-3a11eafe-9e20-11e7-91c0-3a1f7cb2fed9.png"> <img width="1069" alt="screen shot 2017-09-21 at 9 58 10 am" src="https://user-images.githubusercontent.com/25483966/30699739-8be3d0f2-9eb3-11e7-87b1-bcebb03689d1.png"> **History** ![screen shot 2017-09-21 at 10 02 41 am](https://user-images.githubusercontent.com/25483966/30699971-33da138e-9eb4-11e7-9767-871c1225de32.png) **Dependencies** <img width="1104" alt="screen shot 2017-09-21 at 10 05 12 am" src="https://user-images.githubusercontent.com/25483966/30700074-83d095e8-9eb4-11e7-98dc-c3717f580e1a.png"> **Create Group** <img width="553" alt="screen shot 2017-09-21 at 10 56 50 am" src="https://user-images.githubusercontent.com/25483966/30702737-b90abb7e-9ebb-11e7-8563-f0b7926a6b58.png"> **Edit Group** <img width="898" alt="screen shot 2017-09-21 at 10 58 35 am" src="https://user-images.githubusercontent.com/25483966/30702804-f0fc4cf0-9ebb-11e7-9b2e-02a2cafbc2fc.png"> **Audit Logs** <img width="896" alt="screen shot 2017-09-21 at 10 59 54 am" src="https://user-images.githubusercontent.com/25483966/30702910-3bc940b2-9ebc-11e7-9665-6323dae4f35d.png"> **Bulk Operations** <img width="901" alt="screen shot 2017-09-21 at 11 01 50 am" src="https://user-images.githubusercontent.com/25483966/30703000-7a0b3196-9ebc-11e7-9d37-c854f56a1edb.png"> **Create Controller** <img width="353" alt="screen shot 2017-09-21 at 1 13 03 pm" src="https://user-images.githubusercontent.com/25483966/30709014-c3362710-9ece-11e7-8771-c3bb5ed93b7e.png"> **Create Template** <img width="346" alt="screen shot 2017-09-21 at 1 15 57 pm" src="https://user-images.githubusercontent.com/25483966/30709128-239c217c-9ecf-11e7-86b3-a572063150dd.png"> **Context Nav - Users, Sites, Help (in Preview/Dashboard)** <img width="1188" alt="screen shot 2017-09-21 at 1 17 18 pm" src="https://user-images.githubusercontent.com/25483966/30709228-8e8a2240-9ecf-11e7-8493-c9afe76730a5.png"> **Context Nav - Publishing Status, Preview Tools, In-Context Edit, Search (in Preview)** <img width="1148" alt="screen shot 2017-09-21 at 1 21 08 pm" src="https://user-images.githubusercontent.com/25483966/30709333-da8ac6fe-9ecf-11e7-9eae-ca2a7bcfd2c6.png"> There may be some other items not translated, but the above are what I have found so far.
1.0
[studio-ui] Some items in various dialogs are not translated when the language is changed - Some items in various dialogs are not translated when the language is changed. Using the website_editorial bp, change the language by clicking on your user name at the top right of Studio, then select **Settings** Under **Profile**, change the language from **English** to **Spanish**, then click on the **Update Profile** button. Here are the dialogs containing items that may need translation: **Account Management** ![screen shot 2017-09-20 at 10 42 44 am](https://user-images.githubusercontent.com/25483966/30650411-cc795204-9df0-11e7-8a22-5aede16ef674.png) **My Sites** ![screen shot 2017-09-20 at 10 45 40 am](https://user-images.githubusercontent.com/25483966/30650477-fb34e0f4-9df0-11e7-8d80-f79e007c3a14.png) **Dashboard Icon Guide** ![screen shot 2017-09-20 at 10 48 23 am](https://user-images.githubusercontent.com/25483966/30650729-af07d2b2-9df1-11e7-8283-a50ce072e913.png) **About** ![screen shot 2017-09-20 at 10 52 32 am](https://user-images.githubusercontent.com/25483966/30650838-e9eda122-9df1-11e7-91f8-b641a4f77a76.png) **Users** ![screen shot 2017-09-20 at 12 09 45 pm](https://user-images.githubusercontent.com/25483966/30654735-c6488f24-9dfc-11e7-940a-7151612ae53e.png) **Create User** ![screen shot 2017-09-20 at 12 07 37 pm](https://user-images.githubusercontent.com/25483966/30654635-7a2b73d6-9dfc-11e7-90fb-1a0d2fac7827.png) **Publish Status** <img width="373" alt="screen shot 2017-09-20 at 4 12 16 pm" src="https://user-images.githubusercontent.com/25483966/30665374-c798e2a8-9e1e-11e7-82da-0531197f98be.png"> **Targeting** <img width="728" alt="screen shot 2017-09-20 at 4 15 29 pm" src="https://user-images.githubusercontent.com/25483966/30665500-2c4741cc-9e1f-11e7-8cdb-de7ed9b92a03.png"> **Edit Form** <img width="1131" alt="screen shot 2017-09-20 at 4 19 44 pm" src="https://user-images.githubusercontent.com/25483966/30665792-3a11eafe-9e20-11e7-91c0-3a1f7cb2fed9.png"> <img width="1069" alt="screen shot 2017-09-21 at 9 58 10 am" src="https://user-images.githubusercontent.com/25483966/30699739-8be3d0f2-9eb3-11e7-87b1-bcebb03689d1.png"> **History** ![screen shot 2017-09-21 at 10 02 41 am](https://user-images.githubusercontent.com/25483966/30699971-33da138e-9eb4-11e7-9767-871c1225de32.png) **Dependencies** <img width="1104" alt="screen shot 2017-09-21 at 10 05 12 am" src="https://user-images.githubusercontent.com/25483966/30700074-83d095e8-9eb4-11e7-98dc-c3717f580e1a.png"> **Create Group** <img width="553" alt="screen shot 2017-09-21 at 10 56 50 am" src="https://user-images.githubusercontent.com/25483966/30702737-b90abb7e-9ebb-11e7-8563-f0b7926a6b58.png"> **Edit Group** <img width="898" alt="screen shot 2017-09-21 at 10 58 35 am" src="https://user-images.githubusercontent.com/25483966/30702804-f0fc4cf0-9ebb-11e7-9b2e-02a2cafbc2fc.png"> **Audit Logs** <img width="896" alt="screen shot 2017-09-21 at 10 59 54 am" src="https://user-images.githubusercontent.com/25483966/30702910-3bc940b2-9ebc-11e7-9665-6323dae4f35d.png"> **Bulk Operations** <img width="901" alt="screen shot 2017-09-21 at 11 01 50 am" src="https://user-images.githubusercontent.com/25483966/30703000-7a0b3196-9ebc-11e7-9d37-c854f56a1edb.png"> **Create Controller** <img width="353" alt="screen shot 2017-09-21 at 1 13 03 pm" src="https://user-images.githubusercontent.com/25483966/30709014-c3362710-9ece-11e7-8771-c3bb5ed93b7e.png"> **Create Template** <img width="346" alt="screen shot 2017-09-21 at 1 15 57 pm" src="https://user-images.githubusercontent.com/25483966/30709128-239c217c-9ecf-11e7-86b3-a572063150dd.png"> **Context Nav - Users, Sites, Help (in Preview/Dashboard)** <img width="1188" alt="screen shot 2017-09-21 at 1 17 18 pm" src="https://user-images.githubusercontent.com/25483966/30709228-8e8a2240-9ecf-11e7-8493-c9afe76730a5.png"> **Context Nav - Publishing Status, Preview Tools, In-Context Edit, Search (in Preview)** <img width="1148" alt="screen shot 2017-09-21 at 1 21 08 pm" src="https://user-images.githubusercontent.com/25483966/30709333-da8ac6fe-9ecf-11e7-9eae-ca2a7bcfd2c6.png"> There may be some other items not translated, but the above are what I have found so far.
priority
some items in various dialogs are not translated when the language is changed some items in various dialogs are not translated when the language is changed using the website editorial bp change the language by clicking on your user name at the top right of studio then select settings under profile change the language from english to spanish then click on the update profile button here are the dialogs containing items that may need translation account management my sites dashboard icon guide about users create user publish status img width alt screen shot at pm src targeting img width alt screen shot at pm src edit form img width alt screen shot at pm src img width alt screen shot at am src history dependencies img width alt screen shot at am src create group img width alt screen shot at am src edit group img width alt screen shot at am src audit logs img width alt screen shot at am src bulk operations img width alt screen shot at am src create controller img width alt screen shot at pm src create template img width alt screen shot at pm src context nav users sites help in preview dashboard img width alt screen shot at pm src context nav publishing status preview tools in context edit search in preview img width alt screen shot at pm src there may be some other items not translated but the above are what i have found so far
1
305,272
26,374,613,551
IssuesEvent
2023-01-12 00:35:55
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[BACKEND] [REMOTO] Engenheiro de Software Backend na [STONE]
BACK-END BANCO DE DADOS MYSQL SQL SERVER POSTGRESQL REMOTO GITHUB CI/CD APIs TESTES AUTOMATIZADOS DESENVOLVIMENTO WEB HELP WANTED Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - O Time: - O time de Pricing é responsável por precificar os produtos da StoneCo, garantindo a competitividade da companhia assim como a lucratividade das propostas. - O Desafio: - Procuramos um desenvolvedor(a) apaixonado(a) por desafios, tecnologia e conhecimento. Será responsável por, entre outros: - Atuar no desenvolvimento e manutenção de novas funcionalidades - Ter responsabilidade para desenvolver em sistemas de importância crítica para a companhia - Monitorar o desempenho e a aplicação das alterações dos sistemas - Documentar as mudanças realizadas - Aplicar sistemas em produção levando em consideração: - Desempenho das aplicações; - Escalabilidade; - Segurança; - Disponibilidade dos serviços/confiabilidade; - Integração com outros sistemas da empresa (atual ou futura); ## Local - Remoto ## Benefícios - Familiar: Convênio SESC (gratuito); plano de saúde e odontológico (Bradesco TNQ2 sem co-participação) com Concierge. - Filhos: Auxílio-creche para crianças até 5 anos. - Bem-estar: Gympass (bora malhar!!!); horário flexível; trabalho remoto; dress code informal; seguro de vida; vale Transporte (se quiser, é opcional); vale refeição e/ou vale Alimentação e bicicletário - Mentoria e Engajamento:Super incentivamos idas a eventos e treinamentos, e se fizer sentido na estratégia, melhor ainda (qualquer um pode pedir); trabalhar com referências técnicas; eventos internos (hackathon, black friday,confraternizações das tribos, datas comemorativas) e aprendizado contínuo. ## Requisitos **Obrigatórios:** - Autonomia - Gostar de trabalhar em equipe; - Disposição para se adaptar e se aperfeiçoar; - Conhecimento de algoritmos e estruturas de dados; - Conhecimento de pelo menos uma linguagem de programação; - O que é esperado tecnicamente? - Saiba questionar gaps na arquitetura e no processo de desenvolvimento e propor soluções; - Saiba realizar inspeções no código e na documentação; - Implemente testes automatizados; - Experiência com desenvolvimento de APIs e arquitetura de aplicações Web; - Familiaridade com as práticas de Entrega Contínua (CI/CD); - Conhecimento de Bancos de dados relacionais (postgreSQL, MySQL, SQL Server); - Capacidade de colaborar com uma equipe usando GitHub ou similar (Pull Requests, Code Review, Issues); - Cabeça aberta para trabalhar com qualquer tecnologia; **Desejáveis:** - Conhecimentos/Experiência em: - Docker; - React; - Análise de dados; - Desenvolvimento e Manutenção de banco de dados; - Sistemas e Produtos Cloud ( Azure, Google Cloud, AWS); - Linguagens: TypeScript, C#, Java, Python; ## Contratação - a combinar ## Nossa empresa - Aqui na Stech, prezamos pela liberdade e autonomia. Liberdade para escolher seus horários, suas stacks, a forma como você vai trabalhar e mais importante, liberdade de ser quem você é. Você vai ter autonomia para fazer as suas escolhas e será cobrado por elas. Somos todos adultos e gostamos muito de um termo chamado “Accountability”, que podemos traduzir para responsabilidade pelos seus resultados. Queremos que você procure primeiro entender, para depois ser entendido. Mas não se preocupe, você sempre será ouvido, sua opinião conta e conta muito. Por último, e com muita importância para nós: Nós prezamos pela troca de conhecimento. “Be humble and kind”, esteja sempre de coração aberto para aprender novas formas de pensar e conhecimentos diferentes. Queremos achar em vocês inteligência, energia e integridade. E em troca, oferecemos autonomia, liberdade e flexibilidade. ## Como se candidatar - [Clique aqui para se candidatar](https://boards.greenhouse.io/stone/jobs/4185161003?gh_src=c9fa16233us)
1.0
[BACKEND] [REMOTO] Engenheiro de Software Backend na [STONE] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - O Time: - O time de Pricing é responsável por precificar os produtos da StoneCo, garantindo a competitividade da companhia assim como a lucratividade das propostas. - O Desafio: - Procuramos um desenvolvedor(a) apaixonado(a) por desafios, tecnologia e conhecimento. Será responsável por, entre outros: - Atuar no desenvolvimento e manutenção de novas funcionalidades - Ter responsabilidade para desenvolver em sistemas de importância crítica para a companhia - Monitorar o desempenho e a aplicação das alterações dos sistemas - Documentar as mudanças realizadas - Aplicar sistemas em produção levando em consideração: - Desempenho das aplicações; - Escalabilidade; - Segurança; - Disponibilidade dos serviços/confiabilidade; - Integração com outros sistemas da empresa (atual ou futura); ## Local - Remoto ## Benefícios - Familiar: Convênio SESC (gratuito); plano de saúde e odontológico (Bradesco TNQ2 sem co-participação) com Concierge. - Filhos: Auxílio-creche para crianças até 5 anos. - Bem-estar: Gympass (bora malhar!!!); horário flexível; trabalho remoto; dress code informal; seguro de vida; vale Transporte (se quiser, é opcional); vale refeição e/ou vale Alimentação e bicicletário - Mentoria e Engajamento:Super incentivamos idas a eventos e treinamentos, e se fizer sentido na estratégia, melhor ainda (qualquer um pode pedir); trabalhar com referências técnicas; eventos internos (hackathon, black friday,confraternizações das tribos, datas comemorativas) e aprendizado contínuo. ## Requisitos **Obrigatórios:** - Autonomia - Gostar de trabalhar em equipe; - Disposição para se adaptar e se aperfeiçoar; - Conhecimento de algoritmos e estruturas de dados; - Conhecimento de pelo menos uma linguagem de programação; - O que é esperado tecnicamente? - Saiba questionar gaps na arquitetura e no processo de desenvolvimento e propor soluções; - Saiba realizar inspeções no código e na documentação; - Implemente testes automatizados; - Experiência com desenvolvimento de APIs e arquitetura de aplicações Web; - Familiaridade com as práticas de Entrega Contínua (CI/CD); - Conhecimento de Bancos de dados relacionais (postgreSQL, MySQL, SQL Server); - Capacidade de colaborar com uma equipe usando GitHub ou similar (Pull Requests, Code Review, Issues); - Cabeça aberta para trabalhar com qualquer tecnologia; **Desejáveis:** - Conhecimentos/Experiência em: - Docker; - React; - Análise de dados; - Desenvolvimento e Manutenção de banco de dados; - Sistemas e Produtos Cloud ( Azure, Google Cloud, AWS); - Linguagens: TypeScript, C#, Java, Python; ## Contratação - a combinar ## Nossa empresa - Aqui na Stech, prezamos pela liberdade e autonomia. Liberdade para escolher seus horários, suas stacks, a forma como você vai trabalhar e mais importante, liberdade de ser quem você é. Você vai ter autonomia para fazer as suas escolhas e será cobrado por elas. Somos todos adultos e gostamos muito de um termo chamado “Accountability”, que podemos traduzir para responsabilidade pelos seus resultados. Queremos que você procure primeiro entender, para depois ser entendido. Mas não se preocupe, você sempre será ouvido, sua opinião conta e conta muito. Por último, e com muita importância para nós: Nós prezamos pela troca de conhecimento. “Be humble and kind”, esteja sempre de coração aberto para aprender novas formas de pensar e conhecimentos diferentes. Queremos achar em vocês inteligência, energia e integridade. E em troca, oferecemos autonomia, liberdade e flexibilidade. ## Como se candidatar - [Clique aqui para se candidatar](https://boards.greenhouse.io/stone/jobs/4185161003?gh_src=c9fa16233us)
non_priority
engenheiro de software backend na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga o time o time de pricing é responsável por precificar os produtos da stoneco garantindo a competitividade da companhia assim como a lucratividade das propostas o desafio procuramos um desenvolvedor a apaixonado a por desafios tecnologia e conhecimento será responsável por entre outros atuar no desenvolvimento e manutenção de novas funcionalidades ter responsabilidade para desenvolver em sistemas de importância crítica para a companhia monitorar o desempenho e a aplicação das alterações dos sistemas documentar as mudanças realizadas aplicar sistemas em produção levando em consideração desempenho das aplicações escalabilidade segurança disponibilidade dos serviços confiabilidade integração com outros sistemas da empresa atual ou futura local remoto benefícios familiar convênio sesc gratuito plano de saúde e odontológico bradesco sem co participação com concierge filhos auxílio creche para crianças até anos bem estar gympass bora malhar horário flexível trabalho remoto dress code informal seguro de vida vale transporte se quiser é opcional vale refeição e ou vale alimentação e bicicletário mentoria e engajamento super incentivamos idas a eventos e treinamentos e se fizer sentido na estratégia melhor ainda qualquer um pode pedir trabalhar com referências técnicas eventos internos hackathon black friday confraternizações das tribos datas comemorativas e aprendizado contínuo requisitos obrigatórios autonomia gostar de trabalhar em equipe disposição para se adaptar e se aperfeiçoar conhecimento de algoritmos e estruturas de dados conhecimento de pelo menos uma linguagem de programação o que é esperado tecnicamente saiba questionar gaps na arquitetura e no processo de desenvolvimento e propor soluções saiba realizar inspeções no código e na documentação implemente testes automatizados experiência com desenvolvimento de apis e arquitetura de aplicações web familiaridade com as práticas de entrega contínua ci cd conhecimento de bancos de dados relacionais postgresql mysql sql server capacidade de colaborar com uma equipe usando github ou similar pull requests code review issues cabeça aberta para trabalhar com qualquer tecnologia desejáveis conhecimentos experiência em docker react análise de dados desenvolvimento e manutenção de banco de dados sistemas e produtos cloud azure google cloud aws linguagens typescript c java python contratação a combinar nossa empresa aqui na stech prezamos pela liberdade e autonomia liberdade para escolher seus horários suas stacks a forma como você vai trabalhar e mais importante liberdade de ser quem você é você vai ter autonomia para fazer as suas escolhas e será cobrado por elas somos todos adultos e gostamos muito de um termo chamado “accountability” que podemos traduzir para responsabilidade pelos seus resultados queremos que você procure primeiro entender para depois ser entendido mas não se preocupe você sempre será ouvido sua opinião conta e conta muito por último e com muita importância para nós nós prezamos pela troca de conhecimento “be humble and kind” esteja sempre de coração aberto para aprender novas formas de pensar e conhecimentos diferentes queremos achar em vocês inteligência energia e integridade e em troca oferecemos autonomia liberdade e flexibilidade como se candidatar
0
69,353
13,237,596,888
IssuesEvent
2020-08-18 22:02:00
nmrih/source-game
https://api.github.com/repos/nmrih/source-game
closed
(public) kills with "new" variants of weapons don't count in achievement
Priority: Normal Status: Assigned Type: Code
killing with fa_sks: ![nms_favela0004](https://user-images.githubusercontent.com/11333747/37902334-56dce634-311e-11e8-8cc9-5a1ac3a9c433.jpg) killing with fa_sks_nobayo ![nms_favela0005](https://user-images.githubusercontent.com/11333747/37902348-621c878e-311e-11e8-912d-263d11336361.jpg) same with fa_sako85_ironsights, fa_1022_25mag and fa_m16a4_carryhandle
1.0
(public) kills with "new" variants of weapons don't count in achievement - killing with fa_sks: ![nms_favela0004](https://user-images.githubusercontent.com/11333747/37902334-56dce634-311e-11e8-8cc9-5a1ac3a9c433.jpg) killing with fa_sks_nobayo ![nms_favela0005](https://user-images.githubusercontent.com/11333747/37902348-621c878e-311e-11e8-912d-263d11336361.jpg) same with fa_sako85_ironsights, fa_1022_25mag and fa_m16a4_carryhandle
non_priority
public kills with new variants of weapons don t count in achievement killing with fa sks killing with fa sks nobayo same with fa ironsights fa and fa carryhandle
0
110,549
13,913,422,376
IssuesEvent
2020-10-20 20:25:18
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[Design] Healthcare: Healthcare application status
design my-va-dashboard my-va-phase-2 vsa-authenticated-exp
## Background This should be done first. See [Design plan](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/identity-personalization/my-va/2.0-redesign/product/LIH-outline-and-timeline.md) **If a user has healthcare, we must include:** - [ ] Form number - [x] Enrollment Status - [x] Application date - [x] Enrollment Date - [x] Learn more about your VA health benefits: https://staging.va.gov/health-care/about-va-health-benefits/#health-about-basic [Current error state](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/health-care/application/va-application/design/mockups/Dashboard_HCA%20errors.png) **If a user does NOT have healthcare, we need to include**: - [x] Messaging on how to apply/enroll - [x] Link to apply for healthcare https://staging.va.gov/health-care/apply/application **If a user has submitted a health care application, but it is either incomplete, rejected, or closed, we need to include**: - [ ] Application status ## Tasks - [x] Wireframes for all states - [x] Visual design explorations for a user with healthcare - [x] Visual design explorations for a user without healthcare - [x] High fidelity mockups for a user with healthcare - [x] High fidelity mockups for a user without healthcare - [ ] Appropriate stakeholder reviews and approvals have been completed
1.0
[Design] Healthcare: Healthcare application status - ## Background This should be done first. See [Design plan](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/identity-personalization/my-va/2.0-redesign/product/LIH-outline-and-timeline.md) **If a user has healthcare, we must include:** - [ ] Form number - [x] Enrollment Status - [x] Application date - [x] Enrollment Date - [x] Learn more about your VA health benefits: https://staging.va.gov/health-care/about-va-health-benefits/#health-about-basic [Current error state](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/health-care/application/va-application/design/mockups/Dashboard_HCA%20errors.png) **If a user does NOT have healthcare, we need to include**: - [x] Messaging on how to apply/enroll - [x] Link to apply for healthcare https://staging.va.gov/health-care/apply/application **If a user has submitted a health care application, but it is either incomplete, rejected, or closed, we need to include**: - [ ] Application status ## Tasks - [x] Wireframes for all states - [x] Visual design explorations for a user with healthcare - [x] Visual design explorations for a user without healthcare - [x] High fidelity mockups for a user with healthcare - [x] High fidelity mockups for a user without healthcare - [ ] Appropriate stakeholder reviews and approvals have been completed
non_priority
healthcare healthcare application status background this should be done first see if a user has healthcare we must include form number enrollment status application date enrollment date learn more about your va health benefits if a user does not have healthcare we need to include messaging on how to apply enroll link to apply for healthcare if a user has submitted a health care application but it is either incomplete rejected or closed we need to include application status tasks wireframes for all states visual design explorations for a user with healthcare visual design explorations for a user without healthcare high fidelity mockups for a user with healthcare high fidelity mockups for a user without healthcare appropriate stakeholder reviews and approvals have been completed
0
28,146
13,543,390,727
IssuesEvent
2020-09-16 18:53:08
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
.NET 5.0 Microbenchmarks Performance Study Report
Discussion Triaged area-Meta tenet-performance tenet-performance-benchmarks tracking
# Goals The main goal of my study was to ensure that **we ship .NET 5.0 without any performance regressions** and validate whether in the near future we can **fully** rely on the regressions auto-filing bot written by @DrewScoggins. My other goal was to get .NET Library Team members involved and keep on growing the performance culture. **#tl;dr The bot is doing a great job in detecting regressions. Most serious regressions have been already fixed, however a few investigations are still in progress.** # Methodology (and how it evolved) In 2018 I had the pleasure to review @AndreyAkinshin ["Pro .NET Benchmarking"](https://aakinshin.net/prodotnetbenchmarking/) book. The "Statistics for Performance Engineers" and "Performance Analysis and Performance Testing" chapters inspired me to [implement](https://github.com/dotnet/performance/pull/165) a small tool called [Results Comparer](https://github.com/dotnet/performance/tree/master/src/tools/ResultsComparer). The tool uses the [Mann-Whitney U](https://en.wikipedia.org/wiki/Mann%E2%80%93Whitney_U_test) statistical test to detect performance regressions in results exported by BenchmarkDotNet. It's being used (or at least it should) as part of our [benchmarking workflow](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md#Preventing-Regressions) to prevent introducing regressions to .NET. In 2019 I was asked by @danmosemsft to verify .NET Core 3.0 performance. Initially, I’ve run all the [microbenchmarks from dotnet/performance repository](https://github.com/dotnet/performance/tree/master/src/benchmarks/micro) using a single machine with dual boot for Windows 10 and Ubuntu 18.04 x64 and used the Results Comparer to find regressions. It very quickly turned out that such a sample was way too small to make sure that we don’t have **any** regressions. Some benchmarks were simply unstable, some architectures like ARM and ARM64 were simply not covered. Other Linux distros and CPU families were also not covered. Then I’ve run the benchmarks on all the PCs, laptops, and VMs that I could access. But I was still missing AMD and ARM results, so I've asked @tannergooding and @BruceForstall for help. @tannergooding has run the benchmarks on all his AMD machines. @BruceForstall has provided me access to a document that explains how to use ARM machines owned by the JIT Team. This turned out to be an invaluable help as I've used these machines many, many times. Including this year during the 5.0 investigation. After having enough samples to cover our matrix of supported OSes and architectures, I’ve built a simple console app on top of `ResultsComparer` (source code available [here](https://github.com/adamsitnik/performance/blob/compareRuntimes/src/tools/ResultsComparer/Program.cs)). The tool uses the very same statistical test to detect regressions, aggregates the results from all different configurations, and sorts them from the biggest regression to the biggest improvement. Such approach allows for very quick identification of regressions of all kinds: * affecting every configuration <details> #### System.Linq.Tests.Perf_Enumerable.FirstWithPredicate_LastElementMatches(input: IOrderedEnumerable) | Result | Base | Diff | Ratio | Operating System | Bit | | ------ | -------:| --------:| -----:| ----------------------- | ----- | | **Slower** | 570.88 | 3069.76 | 0.19 | Windows 10.0.19041.388 | X64 | | **Slower** | 610.20 | 3674.19 | 0.17 | Windows 10.0.18363.959 | X64 | | **Slower** | 598.37 | 3519.26 | 0.17 | Windows 10.0.18363.959 | X64 | | **Slower** | 700.86 | 4238.85 | 0.17 | Windows 10.0.19041.450 | X64 | | **Slower** | 583.19 | 3538.60 | 0.16 | Windows 10.0.19041.450 | X64 | | **Slower** | 546.58 | 3015.23 | 0.18 | Windows 10.0.19042 | X64 | | **Slower** | 665.53 | 3776.10 | 0.18 | Windows 10.0.19041.450 | X64 | | **Slower** | 515.15 | 3162.05 | 0.16 | Windows 10.0.19041.450 | X64 | | **Slower** | 626.94 | 3928.55 | 0.16 | ubuntu 18.04 | X64 | | **Slower** | 630.90 | 4196.01 | 0.15 | manjaro | X64 | | **Slower** | 813.80 | 4605.57 | 0.18 | pop 20.04 | X64 | | **Slower** | 608.59 | 3587.44 | 0.17 | alpine 3.11 | X64 | | **Slower** | 615.67 | 3390.01 | 0.18 | ubuntu 18.04 | X64 | | **Slower** | 2148.33 | 10335.71 | 0.21 | ubuntu 16.04 | Arm64 | | **Slower** | 2183.77 | 10620.53 | 0.21 | ubuntu 16.04 | Arm64 | | **Slower** | 2163.67 | 10815.16 | 0.20 | ubuntu 16.04 | Arm64 | | **Slower** | 1176.33 | 11641.04 | 0.10 | ubuntu 18.04 | Arm64 | | **Slower** | 1550.48 | 5183.74 | 0.30 | ubuntu 20.04 | Arm64 | | **Slower** | 568.67 | 3637.59 | 0.16 | Windows 10.0.18363.959 | X86 | | **Slower** | 664.86 | 4576.24 | 0.15 | Windows 10.0.19041.450 | X86 | | **Slower** | 972.74 | 8054.46 | 0.12 | Windows 10.0.18363.1016 | Arm | | **Slower** | 790.15 | 5171.92 | 0.15 | macOS Catalina 10.15.6 | X64 | | **Slower** | 668.62 | 4153.54 | 0.16 | macOS Catalina 10.15.6 | X64 | | **Slower** | 743.69 | 4727.58 | 0.16 | macOS Mojave 10.14.5 | X64 | </details> * affecting specific OS families (Windows, Unix) <details> #### System.Globalization.Tests.StringSearch.IsPrefix_DifferentFirstChar(Options: (en-US, IgnoreSymbols, False)) | Result | Base | Diff | Ratio | Operating System | Bit | | ------ | --------:| --------:| -----:| ----------------------- | ----- | | **Slower** | 53.24 | 26589.31 | 0.00 | **Windows** 10.0.19041.388 | X64 | | **Slower** | 65.47 | 28371.93 | 0.00 | **Windows** 10.0.18363.959 | X64 | | **Slower** | 63.89 | 27952.39 | 0.00 | **Windows** 10.0.18363.959 | X64 | | **Slower** | 75.24 | 35910.74 | 0.00 | **Windows** 10.0.19041.450 | X64 | | **Slower** | 67.29 | 55198.94 | 0.00 | **Windows** 10.0.19041.450 | X64 | | **Slower** | 58.36 | 31008.73 | 0.00 | **Windows** 10.0.19042 | X64 | | **Slower** | 70.38 | 34632.87 | 0.00 | **Windows** 10.0.19041.450 | X64 | | **Slower** | 58.92 | 27533.16 | 0.00 | **Windows** 10.0.19041.450 | X64 | | Same | 24197.26 | 24316.40 | 1.00 | ubuntu 18.04 | X64 | | Same | 23317.93 | 23585.42 | 0.99 | manjaro | X64 | | Same | 30855.66 | 30176.99 | 1.02 | pop 20.04 | X64 | | Same | 29081.88 | 28590.29 | 1.02 | alpine 3.11 | X64 | | Same | 23929.07 | 23728.33 | 1.01 | ubuntu 18.04 | X64 | | Same | 51918.86 | 51256.87 | 1.01 | ubuntu 16.04 | Arm64 | | Same | 51674.77 | 51693.86 | 1.00 | ubuntu 16.04 | Arm64 | | Same | 51690.93 | 52015.88 | 0.99 | ubuntu 16.04 | Arm64 | | Same | 61071.92 | 43711.17 | 1.40 | ubuntu 18.04 | Arm64 | | Faster | 43870.66 | 26020.13 | 1.69 | ubuntu 20.04 | Arm64 | | **Slower** | 78.42 | 36208.27 | 0.00 | **Windows** 10.0.18363.959 | X86 | | **Slower** | 88.01 | 42312.37 | 0.00 | **Windows** 10.0.19041.450 | X86 | | **Slower** | 104.29 | 57622.86 | 0.00 | **Windows** 10.0.18363.1016 | Arm | | Same | 38089.02 | 40079.68 | 0.95 | macOS Catalina 10.15.6 | X64 | | Same | 32208.09 | 32537.00 | 0.99 | macOS Catalina 10.15.6 | X64 | | Same | 32575.17 | 32782.69 | 0.99 | macOS Mojave 10.14.5 | X64 | </details> * affecting specific Linux distros <details> #### System.Threading.Tests.Perf_CancellationToken.Cancel | Result | Base | Diff | Ratio | Operating System | Bit | | ------ | -------:| -------:| -----:| ----------------------- | ----- | | Same | 116.42 | 120.28 | 0.97 | Windows 10.0.19041.388 | X64 | | Same | 148.25 | 146.53 | 1.01 | Windows 10.0.18363.959 | X64 | | Same | 144.37 | 144.09 | 1.00 | Windows 10.0.18363.959 | X64 | | Same | 154.82 | 151.57 | 1.02 | Windows 10.0.19041.450 | X64 | | Same | 134.57 | 133.40 | 1.01 | Windows 10.0.19041.450 | X64 | | Same | 122.52 | 119.39 | 1.03 | Windows 10.0.19042 | X64 | | Same | 154.48 | 150.92 | 1.02 | Windows 10.0.19041.450 | X64 | | Same | 128.87 | 122.90 | 1.05 | Windows 10.0.19041.450 | X64 | | Same | 169.50 | 168.46 | 1.01 | ubuntu 18.04 | X64 | | Faster | 171.67 | 155.11 | 1.11 | manjaro | X64 | | Same | 179.54 | 175.17 | 1.02 | pop 20.04 | X64 | | **Slower** | 146.39 | 203.94 | 0.72 | **alpine 3.11** | X64 | | Same | 179.39 | 180.75 | 0.99 | ubuntu 18.04 | X64 | | Same | 1068.08 | 1029.35 | 1.04 | ubuntu 16.04 | Arm64 | | Same | 1066.73 | 1056.79 | 1.01 | ubuntu 16.04 | Arm64 | | Same | 1111.72 | 1037.54 | 1.07 | ubuntu 16.04 | Arm64 | | Same | 751.74 | 622.83 | 1.21 | ubuntu 18.04 | Arm64 | | Faster | 675.51 | 318.18 | 2.12 | ubuntu 20.04 | Arm64 | | Same | 258.80 | 257.15 | 1.01 | Windows 10.0.18363.959 | X86 | | Same | 194.61 | 192.96 | 1.01 | Windows 10.0.19041.450 | X86 | | Same | 486.93 | 508.05 | 0.96 | Windows 10.0.18363.1016 | Arm | | Same | 200.25 | 203.78 | 0.98 | macOS Catalina 10.15.6 | X64 | | Same | 168.62 | 163.47 | 1.03 | macOS Catalina 10.15.6 | X64 | | Same | 174.95 | 177.88 | 0.98 | macOS Mojave 10.14.5 | X64 | </details> * affecting specific CPU families <details> #### System.Buffers.Text.Tests.Base64EncodeDecodeInPlaceTests.Base64EncodeInPlace(NumberOfBytes: 200000000) | Result | Base | Diff | Ratio | Operating System | Bit | Processor Name | | ------ | ------------:| ------------:| -----:| ----------------------- | ----- | --------------------------------------------- | | Same | 125616750.00 | 125476550.00 | 1.00 | Windows 10.0.19041.388 | X64 | AMD Ryzen 9 3900X | | Same | 161388400.00 | 156493500.00 | 1.03 | Windows 10.0.18363.959 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 154933500.00 | 154730800.00 | 1.00 | Windows 10.0.18363.959 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 180481800.00 | 180129900.00 | 1.00 | Windows 10.0.19041.450 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | | **Slower** | 161742300.00 | 211160300.00 | 0.77 | Windows 10.0.19041.450 | X64 | Intel Core i7-6700 CPU 3.40GHz (**Skylake**) | | Same | 152928600.00 | 150232700.00 | 1.02 | Windows 10.0.19042 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | | Same | 206708750.00 | 206860050.00 | 1.00 | Windows 10.0.19041.450 | X64 | Intel Core i7-8650U CPU 1.90GHz (Kaby Lake R) | | **Slower** | 140924300.00 | 185228400.00 | 0.76 | Windows 10.0.19041.450 | X64 | Intel Core i7-8700 CPU 3.20GHz (**Coffee Lake**) | | Same | 154948321.00 | 154788579.50 | 1.00 | ubuntu 18.04 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 175860282.50 | 163007313.50 | 1.08 | manjaro | X64 | Intel Core i7-4771 CPU 3.50GHz (Haswell) | | **Slower** | 199713880.00 | 255270486.50 | 0.78 | pop 20.04 | X64 | Intel Core i7-6600U CPU 2.60GHz (**Skylake**) | | Same | 151256100.00 | 168661900.00 | 0.90 | alpine 3.11 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | | Same | 171229200.00 | 165843050.00 | 1.03 | ubuntu 18.04 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | | Same | 503785101.00 | 505992400.50 | 1.00 | ubuntu 16.04 | Arm64 | Unknown processor | | Same | 503901205.00 | 506190175.00 | 1.00 | ubuntu 16.04 | Arm64 | Unknown processor | | Same | 504131772.50 | 506220395.00 | 1.00 | ubuntu 16.04 | Arm64 | Unknown processor | | Same | 473629200.00 | 541631800.00 | 0.87 | ubuntu 18.04 | Arm64 | Unknown processor | | Same | 331381500.00 | 333779500.00 | 0.99 | ubuntu 20.04 | Arm64 | Unknown processor | | Same | 246876150.00 | 247010200.00 | 1.00 | Windows 10.0.18363.959 | X86 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 290036150.00 | 289409500.00 | 1.00 | Windows 10.0.19041.450 | X86 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | | Same | 418007450.00 | 415404450.00 | 1.01 | Windows 10.0.18363.1016 | Arm | Microsoft SQ1 3.0 GHz | | Same | 204196936.50 | 204410652.50 | 1.00 | macOS Catalina 10.15.6 | X64 | Intel Core i5-4278U CPU 2.60GHz (Haswell) | | Same | 176763730.00 | 175647563.50 | 1.01 | macOS Catalina 10.15.6 | X64 | Intel Core i7-4870HQ CPU 2.50GHz (Haswell) | | Same | 180812724.00 | 184849205.00 | 0.98 | macOS Mojave 10.14.5 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | </details> Using the tool had one major flaw: it was not automated and hence we were finding out about the regressions only when we searched for them. This has been recognized and a new project has been started. In 2020 @DrewScoggins started implementing a GitHub bot that would be using the data gathered from performance lab (a set of machines owned by .NET Performance Team) microbenchmark runs to detect and auto-file the regressions. So far the bot was reporting new issues in a [dedicated repository](https://github.com/DrewScoggins/performance-2/issues) and once a week the workgroup led by @DrewScoggins that consisted of @AndyAyersMS, @kunalspathak, @tannergooding any myself was going through the list and triaging the issues. Issues that were seemed as actual regressions were labeled as [Needs Transfer](https://github.com/DrewScoggins/performance-2/issues?q=is%3Aissue+label%3A%22Needs+Transfer%22+) and were later moved by @DrewScoggins to the [runtime repo](https://github.com/dotnet/runtime/issues?q=is%3Aissue+author%3ADrewScoggins+%22%5BPerf%22+). A few weeks ago we were getting close to "code freeze" for .NET 5 and I have asked myself a question: are we sure that the bot has reported all possible regressions for [all the supported OS versions](https://github.com/dotnet/core/blob/master/release-notes/5.0/5.0-supported-os.md)? The bot is using different statistical methods to detect regressions and so far it has been enabled only for Windows 10 x64, Ubuntu 18.04 x64, and Windows 10 x86. So I've decided to spend *some time* and use the old tool that I wrote to verify it. To increase the sample size and get other .NET Libraries Team members involved, I've simply asked the Team to run the benchmarks and share the results with me. Running the performance repo microbenchmarks against the latest .NET Core SDK is super easy thanks to a python script [implemented](https://github.com/dotnet/performance/pull/33) by @jorive. The script downloads the right SDK and starts benchmarking with cleared environment variables. ```cmd git clone https://github.com/dotnet/performance.git python3 ./performance/scripts/benchmarks_ci.py -f netcoreapp3.1 netcoreapp5.0 --filter '*' ``` # Data The data I've received from the .NET Libraries Team members allowed me a big part of the entire [matrix of the supported configurations](https://github.com/dotnet/core/blob/master/release-notes/5.0/5.0-supported-os.md): | Operating System | Arch | Processor Name | Provided by | | ----------------------- | ----- | --------------------------------------------- | ------------------- | | Windows 10.0.19041.388 | X64 | AMD Ryzen 9 3900X | @tannergooding | | Windows 10.0.18363.959 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | @adamsitnik | | Windows 10.0.19041.450 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | @adamsitnik | | Windows 10.0.19041.450 | X64 | Intel Core i7-6700 CPU 3.40GHz (Skylake) | @GrabYourPitchforks | | Windows 10.0.19042 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | @danmosemsft | | Windows 10.0.19041.450 | X64 | Intel Core i7-8650U CPU 1.90GHz (Kaby Lake R) | @jeffhandley | | Windows 10.0.19041.450 | X64 | Intel Core i7-8700 CPU 3.20GHz (Coffee Lake) | @jeffhandley | | ubuntu 18.04 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | @adamsitnik | | manjaro | X64 | Intel Core i7-4771 CPU 3.50GHz (Haswell) | @ManickaP | | pop 20.04 | X64 | Intel Core i7-6600U CPU 2.60GHz (Skylake) | @carlossanlop | | alpine 3.11 (WSL2) | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | @danmosemsft | | ubuntu 18.04 (WSL2) | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | @danmosemsft | | ubuntu 16.04 | Arm64 | Qualcomm Centriq | @adamsitnik | | ubuntu 18.04 (WSL2) | Arm64 | Microsoft SQ1 3.0 GHz (Surface Pro X) | @carlossanlop | | ubuntu 20.04 (WSL2) | Arm64 | Microsoft SQ1 3.0 GHz (Surface Pro X) | @pgovind | | Windows 10.0.18363.959 | X86 | Intel Xeon CPU E5-1650 v4 3.60GHz | @adamsitnik | | Windows 10.0.19041.450 | X86 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | @adamsitnik | | Windows 10.0.18363.1016 | Arm | Microsoft SQ1 3.0 GHz (Surface Pro X) | @adamsitnik | | macOS Catalina 10.15.6 | X64 | Intel Core i5-4278U CPU 2.60GHz (Haswell) | @jeffhandley | | macOS Catalina 10.15.6 | X64 | Intel Core i7-4870HQ CPU 2.50GHz (Haswell) | @carlossanlop | | macOS Mojave 10.14.5 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | @adamsitnik | Everyone interested can download the data from [here](https://microsofteur-my.sharepoint.com/:u:/g/personal/adsitnik_microsoft_com/Ea0pIss74N9GsDywOk3Wq_0BFcyFSEJ6m5xFMseha1T0IA?e=bWQpIp). The full report generated by the tool is available [here](https://microsofteur-my.sharepoint.com/:t:/g/personal/adsitnik_microsoft_com/EY3O7vVrCX9OkulTAowPbToBGFmQrTfosY0LDhXBxkytAQ?e=Aj0BRS). Moreover, **the full historical data turned out to be extremely useful**. I've used it every time I was not sure whether something was a regression or just unstable|multimodal benchmark: * [Windows 10 x64](https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/master_x64_Windows%2010.0.18362/PerfLabTests.BlockCopyPerf.CallBlockCopy(numElements:%201000).html) * [Ubuntu 18.04 x64](https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/09_01_2020/refs/heads/master_x64_ubuntu%2018.04/System.Threading.Tests.Perf_CancellationToken.html) * [Ubuntu 18.04 ARM64](https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/master_arm64_ubuntu%2018.04/AllTestindex.html) (added yesterday) * [Windows 10 x86](https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/master_x86_Windows%2010.0.18362/PerfLabTests.BlockCopyPerf.CallBlockCopy(numElements:%201000).html) # Regressions ## Already fixed - [x] `System.Collections.Contains*`, `System.Memory.SequenceReader.TryReadTo`, `System.Text.Json.Tests.Perf_Segment.ReadSingleSegmentSequenceByN` * was a 32 bit issue only (both `x86` and `ARM`) * detected by the bot, reported in https://github.com/DrewScoggins/performance-2/issues/910#issuecomment-676413559 * confirmed: https://github.com/DrewScoggins/performance-2/issues/910#issuecomment-677854221 * transffered to runtime repo: https://github.com/dotnet/runtime/issues/41167 * fixed in https://github.com/dotnet/runtime/pull/41198 * backported to 5.0 in https://github.com/dotnet/runtime/pull/41254 - [x] `System.Collections.CtorGivenSize<Int32>.Array(Size: 512)` * specific to Apline only * created an issue https://github.com/dotnet/runtime/issues/41398 * confirmed by @jkotas to be not WSL specific, but a much bigger Alpine perf problem * it has shown that an increased number of Gen 0 collections is a valuable metric to detect regressions * fixed in https://github.com/dotnet/runtime/pull/41532 * backported to 5.0 https://github.com/dotnet/runtime/pull/41547 * created https://github.com/dotnet/runtime/issues/41708 to add unit tests that ensure that this problem is not coming back - [x] `System.Numerics.Tests.Perf_Quaternion.Conjugate` and `System.Numerics.Tests.Perf_Quaternion.Negat*` * not reported by the bot because it's a brand new benchmark and we did not have historical data at the time of my investigation * issue created: https://github.com/dotnet/runtime/issues/41738 * fixed in https://github.com/dotnet/runtime/pull/41829 * backported to 5.0-rc2 in https://github.com/dotnet/runtime/pull/41885 - [x] `Directory.EnumerateFiles` * not reported by the bot, most probably because it was a very fresh regression * issue created: https://github.com/dotnet/runtime/issues/41739 * fixed in https://github.com/dotnet/runtime/issues/41739 * backported to 5.0-rc2 in https://github.com/dotnet/runtime/pull/41820 - [x] `ByteMark.BenchIDEAEncryption` * not reported by the bot, most probably because it was a very fresh regression * issue created: https://github.com/dotnet/runtime/issues/41677 * fixed in https://github.com/dotnet/runtime/pull/40871 * backported to 5.0-rc2 in https://github.com/dotnet/runtime/pull/41838 - [x] `System.Text.Perf_Utf8Encoding` * not detected by the bot because it was not enabled for ARM yet * issue created: https://github.com/dotnet/runtime/issues/41699 * fixed in https://github.com/dotnet/runtime/pull/42052 * backported to 5.0-rc2 in #42064 ## Investigation in progress - [ ] `System.Text.Encodings.Web.Tests.Perf_Encoders.EncodeUtf8` * not reported by the bot because it was a brand new benchmark and we did not have historical data at the time of my investigation * issue created: https://github.com/dotnet/runtime/issues/41104 * investigation is in progress - [ ] `System.Memory.Slice` * not detected by the bot because it was not enabled for ARM yet * seems to be ARM64-specific, created an issue https://github.com/dotnet/runtime/issues/41704 * investigation is in progress - [ ] `PerfLabTests.CastingPerf2.CastingPerf.IntObj` * not detected by the bot because it was not enabled for ARM yet * seems to be ARM64-specific, created an issue https://github.com/dotnet/runtime/issues/41706 * investigation is in progress ## By design or Acceptable - [ ] ICU-related regressions * `System.Globalization.Tests.StringSearch`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/37819 * `System.Memory.ReadOnlySpan.IndexOfString`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/39724 * `System.Globalization.Tests.Perf_DateTimeCultureInfo.Parse(culturestring: ja)`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/37807 * `System.Globalization.Tests.StringEquality`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/39038 * I've created one uber issue to track all of them in one place: https://github.com/dotnet/runtime/issues/40942 * `OrdinalIgnoreCase` has been optimized in https://github.com/dotnet/runtime/pull/40962 * TODO: doc update still required - [x] `System.Linq.Tests.Perf_Enumerable.FirstWithPredicate_LastElementMatches(input: IOrderedEnumerable)` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39032 * closed, by design: removed the `O(N log N)` cost of the `OrderBy` https://github.com/dotnet/runtime/issues/39032#issuecomment-656678750 - [x] `System.Collections.Tests.Perf_BitArray.*(Size: 4)` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37813 * closed, by design: introduction of vectorization has increased the cost of operations for small inputs: https://github.com/dotnet/runtime/issues/37813#issuecomment-656370853 - [x] `System.Threading.Tests.Perf_Thread.GetCurrentProcessorId` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37804 * closed, by design: precision was improved at a cost of acceptable minor perf regression: https://github.com/dotnet/runtime/issues/37804#issuecomment-643448336 - [x] `PerfLabTests.CastingPerf.CheckIsInstAnyIsInterfaceNo`, `PerfLabTests.CastingPerf.CheckObjIsInterfaceNo` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37803 * closed, by design: known tradeoff: https://github.com/dotnet/runtime/issues/37803#issuecomment-670209689 - [x] `System.Net.NetworkInformation.Tests.PhysicalAddressTests.PAShort` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39720 * closed, acceptable for improved code reuse https://github.com/dotnet/runtime/issues/39720#issuecomment-671535578 * benchmark for 1 byte removed, added 6 bytes in https://github.com/dotnet/performance/pull/1490 - [x] `System.Numerics.Tests.Perf_Vector*.GetHashCodeBenchmark` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39035 and https://github.com/dotnet/runtime/issues/39029 * closed, "it should not be used" https://github.com/dotnet/runtime/issues/39029#issuecomment-656408464 - [ ] `System.Net.Primitives.Tests.CredentialCacheTests.ForEach(uriCount: 0, hostPortCount: 0)` * detected by the bot, reported in https://github.com/DrewScoggins/performance-2/issues/510 * confirmed: https://github.com/DrewScoggins/performance-2/issues/510#issuecomment-680783031 * awaiting the transfer to runtime repo. Most probably a by-design regression. ## Moved to 6.0 - [x] `System.Tests.Perf_Char.GetUnicodeCategory(c: '?')` * detected and reported by the bot in https://github.com/DrewScoggins/performance-2/issues/574, I've created https://github.com/dotnet/runtime/issues/41107 * minor regression for non-ascii characters, moved to 6.0 - [x] `PerfLabTests.StackWalk.Walk` * detected by the bot and reported in https://github.com/dotnet/runtime/issues/39115 * confirmed in https://github.com/dotnet/runtime/issues/39115#issuecomment-677857078 * specific to everything that is not Windows x64, rather not critical -> moved to 6.0: https://github.com/dotnet/runtime/issues/39115#issuecomment-682684126 - [x] `System.Tests.Perf_String.Replace_Char(text: "Hello", oldChar: 'l', newChar: '!')` * reported in https://github.com/dotnet/runtime/issues/37816 * confirmed in https://github.com/dotnet/runtime/issues/37816#issuecomment-680706314 * moved to 6.0 - [x] `System.Text.Perf_Utf8String.IsAscii(Input: EnglishAllAscii)` * not reported by the bot because it was a brand new benchmark and we did not have historical data at the time of my investigation * issue created: https://github.com/dotnet/runtime/issues/41388 * moved to 6.0 as `Utf8String` is still only experimental ## Unstable or multimodal benchmarks There was of course more of them, here are the ones that I've noted to use as Contract Tests in the near future (to reduce the noise produced by the bot): * `System.Buffers.Tests.RentReturnArrayPoolTests<Byte>.ProducerConsumer` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39031 * asked for historical data to verify if it's multimodal or not https://github.com/dotnet/runtime/issues/39031#issuecomment-680712003 * thanks to historical data provided it was possible to tell that it's unstable for x64 and bimodal for x86: https://github.com/dotnet/runtime/issues/39031#issuecomment-682207270 * `System.Memory.ReadOnlySequence.Slice_Repeat_StartPosition_And_EndPosition(Segment: Multiple)` * quite unstable benchmark, I've verified that 5.0 codegen is better * `PerfLabTests.BlockCopyPerf.CallBlockCopy` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37808 * copying 0 elements does not add value: https://github.com/dotnet/runtime/issues/37808#issuecomment-654424436 * test case for copying 0 elements removed in https://github.com/dotnet/performance/pull/1465 * closed as unstable based on full historical data: https://github.com/dotnet/runtime/issues/37808#issuecomment-685014522 * `System.Tests.Perf_String.Trim_CharArr(s: "Test", c: [' ', ' '])` * multimodal benchmark, needs a rewrite as stated long time ago: https://github.com/dotnet/runtime/issues/13135 * `System.Threading.Tests.Perf_Interlocked.CompareExchange_long` * the benchmark typically reports `10ns`, but sometimes `x100` that. Only for `x86`. I need logs to verify whether it's a BDN bug or not. * issue created https://github.com/dotnet/performance/issues/1497 * `System.Memory.Span<Int32>.IndexOfValue(Size: 512)` * reported in https://github.com/dotnet/runtime/issues/39722 * confirmed that it was due to code alignment change in https://github.com/dotnet/runtime/issues/39722#issuecomment-674999435 * `Benchstone.BenchI.Fib.Test` * perfectly multimodal, great example for a contract test # Summary * The bot has reported all major performance issues for the configurations that it was enabled for (Windows x64, x86, and Ubuntu x64). Great work @DrewScoggins! * The full historical data turned out to be extremely useful to exclude all false positives for multimodal and unstable benchmarks. * We have missed one important x86 bug during triaging (human error), but it got discovered during the study (https://github.com/dotnet/runtime/issues/41167#issuecomment-679285756). To avoid such problems in the future and to enable the bot in the runtime repo, the noise of the bot needs to be reduced. Currently, it's quite high, mostly due to the multimodal nature of the benchmarks. * The study has detected relatively many new ARM64 perf problems at a late stage of the release. The sooner we enable the bot for ARM64, the better. Moreover, we should be more frequently asking for ARM64 results when reviewing big changes that affect the performance of frequently used features (like sorting the arrays). * The study has shown that measuring the performance of `GNU libc` based Linux distros like Ubuntu is not enough to detect `musl libc` specific regressions. We should consider adding Alpine runs to the perf lab. * This time no important issues specific to `macOS` and different CPU families were discovered. It has proven that the perf lab has good hardware coverage. * The Alpine regression has shown that an increased number of Gen 0 collections can be a very valuable metric to detect regressions. We should consider extending the bot to use it. **Big thanks to everyone involved!**
True
.NET 5.0 Microbenchmarks Performance Study Report - # Goals The main goal of my study was to ensure that **we ship .NET 5.0 without any performance regressions** and validate whether in the near future we can **fully** rely on the regressions auto-filing bot written by @DrewScoggins. My other goal was to get .NET Library Team members involved and keep on growing the performance culture. **#tl;dr The bot is doing a great job in detecting regressions. Most serious regressions have been already fixed, however a few investigations are still in progress.** # Methodology (and how it evolved) In 2018 I had the pleasure to review @AndreyAkinshin ["Pro .NET Benchmarking"](https://aakinshin.net/prodotnetbenchmarking/) book. The "Statistics for Performance Engineers" and "Performance Analysis and Performance Testing" chapters inspired me to [implement](https://github.com/dotnet/performance/pull/165) a small tool called [Results Comparer](https://github.com/dotnet/performance/tree/master/src/tools/ResultsComparer). The tool uses the [Mann-Whitney U](https://en.wikipedia.org/wiki/Mann%E2%80%93Whitney_U_test) statistical test to detect performance regressions in results exported by BenchmarkDotNet. It's being used (or at least it should) as part of our [benchmarking workflow](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md#Preventing-Regressions) to prevent introducing regressions to .NET. In 2019 I was asked by @danmosemsft to verify .NET Core 3.0 performance. Initially, I’ve run all the [microbenchmarks from dotnet/performance repository](https://github.com/dotnet/performance/tree/master/src/benchmarks/micro) using a single machine with dual boot for Windows 10 and Ubuntu 18.04 x64 and used the Results Comparer to find regressions. It very quickly turned out that such a sample was way too small to make sure that we don’t have **any** regressions. Some benchmarks were simply unstable, some architectures like ARM and ARM64 were simply not covered. Other Linux distros and CPU families were also not covered. Then I’ve run the benchmarks on all the PCs, laptops, and VMs that I could access. But I was still missing AMD and ARM results, so I've asked @tannergooding and @BruceForstall for help. @tannergooding has run the benchmarks on all his AMD machines. @BruceForstall has provided me access to a document that explains how to use ARM machines owned by the JIT Team. This turned out to be an invaluable help as I've used these machines many, many times. Including this year during the 5.0 investigation. After having enough samples to cover our matrix of supported OSes and architectures, I’ve built a simple console app on top of `ResultsComparer` (source code available [here](https://github.com/adamsitnik/performance/blob/compareRuntimes/src/tools/ResultsComparer/Program.cs)). The tool uses the very same statistical test to detect regressions, aggregates the results from all different configurations, and sorts them from the biggest regression to the biggest improvement. Such approach allows for very quick identification of regressions of all kinds: * affecting every configuration <details> #### System.Linq.Tests.Perf_Enumerable.FirstWithPredicate_LastElementMatches(input: IOrderedEnumerable) | Result | Base | Diff | Ratio | Operating System | Bit | | ------ | -------:| --------:| -----:| ----------------------- | ----- | | **Slower** | 570.88 | 3069.76 | 0.19 | Windows 10.0.19041.388 | X64 | | **Slower** | 610.20 | 3674.19 | 0.17 | Windows 10.0.18363.959 | X64 | | **Slower** | 598.37 | 3519.26 | 0.17 | Windows 10.0.18363.959 | X64 | | **Slower** | 700.86 | 4238.85 | 0.17 | Windows 10.0.19041.450 | X64 | | **Slower** | 583.19 | 3538.60 | 0.16 | Windows 10.0.19041.450 | X64 | | **Slower** | 546.58 | 3015.23 | 0.18 | Windows 10.0.19042 | X64 | | **Slower** | 665.53 | 3776.10 | 0.18 | Windows 10.0.19041.450 | X64 | | **Slower** | 515.15 | 3162.05 | 0.16 | Windows 10.0.19041.450 | X64 | | **Slower** | 626.94 | 3928.55 | 0.16 | ubuntu 18.04 | X64 | | **Slower** | 630.90 | 4196.01 | 0.15 | manjaro | X64 | | **Slower** | 813.80 | 4605.57 | 0.18 | pop 20.04 | X64 | | **Slower** | 608.59 | 3587.44 | 0.17 | alpine 3.11 | X64 | | **Slower** | 615.67 | 3390.01 | 0.18 | ubuntu 18.04 | X64 | | **Slower** | 2148.33 | 10335.71 | 0.21 | ubuntu 16.04 | Arm64 | | **Slower** | 2183.77 | 10620.53 | 0.21 | ubuntu 16.04 | Arm64 | | **Slower** | 2163.67 | 10815.16 | 0.20 | ubuntu 16.04 | Arm64 | | **Slower** | 1176.33 | 11641.04 | 0.10 | ubuntu 18.04 | Arm64 | | **Slower** | 1550.48 | 5183.74 | 0.30 | ubuntu 20.04 | Arm64 | | **Slower** | 568.67 | 3637.59 | 0.16 | Windows 10.0.18363.959 | X86 | | **Slower** | 664.86 | 4576.24 | 0.15 | Windows 10.0.19041.450 | X86 | | **Slower** | 972.74 | 8054.46 | 0.12 | Windows 10.0.18363.1016 | Arm | | **Slower** | 790.15 | 5171.92 | 0.15 | macOS Catalina 10.15.6 | X64 | | **Slower** | 668.62 | 4153.54 | 0.16 | macOS Catalina 10.15.6 | X64 | | **Slower** | 743.69 | 4727.58 | 0.16 | macOS Mojave 10.14.5 | X64 | </details> * affecting specific OS families (Windows, Unix) <details> #### System.Globalization.Tests.StringSearch.IsPrefix_DifferentFirstChar(Options: (en-US, IgnoreSymbols, False)) | Result | Base | Diff | Ratio | Operating System | Bit | | ------ | --------:| --------:| -----:| ----------------------- | ----- | | **Slower** | 53.24 | 26589.31 | 0.00 | **Windows** 10.0.19041.388 | X64 | | **Slower** | 65.47 | 28371.93 | 0.00 | **Windows** 10.0.18363.959 | X64 | | **Slower** | 63.89 | 27952.39 | 0.00 | **Windows** 10.0.18363.959 | X64 | | **Slower** | 75.24 | 35910.74 | 0.00 | **Windows** 10.0.19041.450 | X64 | | **Slower** | 67.29 | 55198.94 | 0.00 | **Windows** 10.0.19041.450 | X64 | | **Slower** | 58.36 | 31008.73 | 0.00 | **Windows** 10.0.19042 | X64 | | **Slower** | 70.38 | 34632.87 | 0.00 | **Windows** 10.0.19041.450 | X64 | | **Slower** | 58.92 | 27533.16 | 0.00 | **Windows** 10.0.19041.450 | X64 | | Same | 24197.26 | 24316.40 | 1.00 | ubuntu 18.04 | X64 | | Same | 23317.93 | 23585.42 | 0.99 | manjaro | X64 | | Same | 30855.66 | 30176.99 | 1.02 | pop 20.04 | X64 | | Same | 29081.88 | 28590.29 | 1.02 | alpine 3.11 | X64 | | Same | 23929.07 | 23728.33 | 1.01 | ubuntu 18.04 | X64 | | Same | 51918.86 | 51256.87 | 1.01 | ubuntu 16.04 | Arm64 | | Same | 51674.77 | 51693.86 | 1.00 | ubuntu 16.04 | Arm64 | | Same | 51690.93 | 52015.88 | 0.99 | ubuntu 16.04 | Arm64 | | Same | 61071.92 | 43711.17 | 1.40 | ubuntu 18.04 | Arm64 | | Faster | 43870.66 | 26020.13 | 1.69 | ubuntu 20.04 | Arm64 | | **Slower** | 78.42 | 36208.27 | 0.00 | **Windows** 10.0.18363.959 | X86 | | **Slower** | 88.01 | 42312.37 | 0.00 | **Windows** 10.0.19041.450 | X86 | | **Slower** | 104.29 | 57622.86 | 0.00 | **Windows** 10.0.18363.1016 | Arm | | Same | 38089.02 | 40079.68 | 0.95 | macOS Catalina 10.15.6 | X64 | | Same | 32208.09 | 32537.00 | 0.99 | macOS Catalina 10.15.6 | X64 | | Same | 32575.17 | 32782.69 | 0.99 | macOS Mojave 10.14.5 | X64 | </details> * affecting specific Linux distros <details> #### System.Threading.Tests.Perf_CancellationToken.Cancel | Result | Base | Diff | Ratio | Operating System | Bit | | ------ | -------:| -------:| -----:| ----------------------- | ----- | | Same | 116.42 | 120.28 | 0.97 | Windows 10.0.19041.388 | X64 | | Same | 148.25 | 146.53 | 1.01 | Windows 10.0.18363.959 | X64 | | Same | 144.37 | 144.09 | 1.00 | Windows 10.0.18363.959 | X64 | | Same | 154.82 | 151.57 | 1.02 | Windows 10.0.19041.450 | X64 | | Same | 134.57 | 133.40 | 1.01 | Windows 10.0.19041.450 | X64 | | Same | 122.52 | 119.39 | 1.03 | Windows 10.0.19042 | X64 | | Same | 154.48 | 150.92 | 1.02 | Windows 10.0.19041.450 | X64 | | Same | 128.87 | 122.90 | 1.05 | Windows 10.0.19041.450 | X64 | | Same | 169.50 | 168.46 | 1.01 | ubuntu 18.04 | X64 | | Faster | 171.67 | 155.11 | 1.11 | manjaro | X64 | | Same | 179.54 | 175.17 | 1.02 | pop 20.04 | X64 | | **Slower** | 146.39 | 203.94 | 0.72 | **alpine 3.11** | X64 | | Same | 179.39 | 180.75 | 0.99 | ubuntu 18.04 | X64 | | Same | 1068.08 | 1029.35 | 1.04 | ubuntu 16.04 | Arm64 | | Same | 1066.73 | 1056.79 | 1.01 | ubuntu 16.04 | Arm64 | | Same | 1111.72 | 1037.54 | 1.07 | ubuntu 16.04 | Arm64 | | Same | 751.74 | 622.83 | 1.21 | ubuntu 18.04 | Arm64 | | Faster | 675.51 | 318.18 | 2.12 | ubuntu 20.04 | Arm64 | | Same | 258.80 | 257.15 | 1.01 | Windows 10.0.18363.959 | X86 | | Same | 194.61 | 192.96 | 1.01 | Windows 10.0.19041.450 | X86 | | Same | 486.93 | 508.05 | 0.96 | Windows 10.0.18363.1016 | Arm | | Same | 200.25 | 203.78 | 0.98 | macOS Catalina 10.15.6 | X64 | | Same | 168.62 | 163.47 | 1.03 | macOS Catalina 10.15.6 | X64 | | Same | 174.95 | 177.88 | 0.98 | macOS Mojave 10.14.5 | X64 | </details> * affecting specific CPU families <details> #### System.Buffers.Text.Tests.Base64EncodeDecodeInPlaceTests.Base64EncodeInPlace(NumberOfBytes: 200000000) | Result | Base | Diff | Ratio | Operating System | Bit | Processor Name | | ------ | ------------:| ------------:| -----:| ----------------------- | ----- | --------------------------------------------- | | Same | 125616750.00 | 125476550.00 | 1.00 | Windows 10.0.19041.388 | X64 | AMD Ryzen 9 3900X | | Same | 161388400.00 | 156493500.00 | 1.03 | Windows 10.0.18363.959 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 154933500.00 | 154730800.00 | 1.00 | Windows 10.0.18363.959 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 180481800.00 | 180129900.00 | 1.00 | Windows 10.0.19041.450 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | | **Slower** | 161742300.00 | 211160300.00 | 0.77 | Windows 10.0.19041.450 | X64 | Intel Core i7-6700 CPU 3.40GHz (**Skylake**) | | Same | 152928600.00 | 150232700.00 | 1.02 | Windows 10.0.19042 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | | Same | 206708750.00 | 206860050.00 | 1.00 | Windows 10.0.19041.450 | X64 | Intel Core i7-8650U CPU 1.90GHz (Kaby Lake R) | | **Slower** | 140924300.00 | 185228400.00 | 0.76 | Windows 10.0.19041.450 | X64 | Intel Core i7-8700 CPU 3.20GHz (**Coffee Lake**) | | Same | 154948321.00 | 154788579.50 | 1.00 | ubuntu 18.04 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 175860282.50 | 163007313.50 | 1.08 | manjaro | X64 | Intel Core i7-4771 CPU 3.50GHz (Haswell) | | **Slower** | 199713880.00 | 255270486.50 | 0.78 | pop 20.04 | X64 | Intel Core i7-6600U CPU 2.60GHz (**Skylake**) | | Same | 151256100.00 | 168661900.00 | 0.90 | alpine 3.11 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | | Same | 171229200.00 | 165843050.00 | 1.03 | ubuntu 18.04 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | | Same | 503785101.00 | 505992400.50 | 1.00 | ubuntu 16.04 | Arm64 | Unknown processor | | Same | 503901205.00 | 506190175.00 | 1.00 | ubuntu 16.04 | Arm64 | Unknown processor | | Same | 504131772.50 | 506220395.00 | 1.00 | ubuntu 16.04 | Arm64 | Unknown processor | | Same | 473629200.00 | 541631800.00 | 0.87 | ubuntu 18.04 | Arm64 | Unknown processor | | Same | 331381500.00 | 333779500.00 | 0.99 | ubuntu 20.04 | Arm64 | Unknown processor | | Same | 246876150.00 | 247010200.00 | 1.00 | Windows 10.0.18363.959 | X86 | Intel Xeon CPU E5-1650 v4 3.60GHz | | Same | 290036150.00 | 289409500.00 | 1.00 | Windows 10.0.19041.450 | X86 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | | Same | 418007450.00 | 415404450.00 | 1.01 | Windows 10.0.18363.1016 | Arm | Microsoft SQ1 3.0 GHz | | Same | 204196936.50 | 204410652.50 | 1.00 | macOS Catalina 10.15.6 | X64 | Intel Core i5-4278U CPU 2.60GHz (Haswell) | | Same | 176763730.00 | 175647563.50 | 1.01 | macOS Catalina 10.15.6 | X64 | Intel Core i7-4870HQ CPU 2.50GHz (Haswell) | | Same | 180812724.00 | 184849205.00 | 0.98 | macOS Mojave 10.14.5 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | </details> Using the tool had one major flaw: it was not automated and hence we were finding out about the regressions only when we searched for them. This has been recognized and a new project has been started. In 2020 @DrewScoggins started implementing a GitHub bot that would be using the data gathered from performance lab (a set of machines owned by .NET Performance Team) microbenchmark runs to detect and auto-file the regressions. So far the bot was reporting new issues in a [dedicated repository](https://github.com/DrewScoggins/performance-2/issues) and once a week the workgroup led by @DrewScoggins that consisted of @AndyAyersMS, @kunalspathak, @tannergooding any myself was going through the list and triaging the issues. Issues that were seemed as actual regressions were labeled as [Needs Transfer](https://github.com/DrewScoggins/performance-2/issues?q=is%3Aissue+label%3A%22Needs+Transfer%22+) and were later moved by @DrewScoggins to the [runtime repo](https://github.com/dotnet/runtime/issues?q=is%3Aissue+author%3ADrewScoggins+%22%5BPerf%22+). A few weeks ago we were getting close to "code freeze" for .NET 5 and I have asked myself a question: are we sure that the bot has reported all possible regressions for [all the supported OS versions](https://github.com/dotnet/core/blob/master/release-notes/5.0/5.0-supported-os.md)? The bot is using different statistical methods to detect regressions and so far it has been enabled only for Windows 10 x64, Ubuntu 18.04 x64, and Windows 10 x86. So I've decided to spend *some time* and use the old tool that I wrote to verify it. To increase the sample size and get other .NET Libraries Team members involved, I've simply asked the Team to run the benchmarks and share the results with me. Running the performance repo microbenchmarks against the latest .NET Core SDK is super easy thanks to a python script [implemented](https://github.com/dotnet/performance/pull/33) by @jorive. The script downloads the right SDK and starts benchmarking with cleared environment variables. ```cmd git clone https://github.com/dotnet/performance.git python3 ./performance/scripts/benchmarks_ci.py -f netcoreapp3.1 netcoreapp5.0 --filter '*' ``` # Data The data I've received from the .NET Libraries Team members allowed me a big part of the entire [matrix of the supported configurations](https://github.com/dotnet/core/blob/master/release-notes/5.0/5.0-supported-os.md): | Operating System | Arch | Processor Name | Provided by | | ----------------------- | ----- | --------------------------------------------- | ------------------- | | Windows 10.0.19041.388 | X64 | AMD Ryzen 9 3900X | @tannergooding | | Windows 10.0.18363.959 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | @adamsitnik | | Windows 10.0.19041.450 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | @adamsitnik | | Windows 10.0.19041.450 | X64 | Intel Core i7-6700 CPU 3.40GHz (Skylake) | @GrabYourPitchforks | | Windows 10.0.19042 | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | @danmosemsft | | Windows 10.0.19041.450 | X64 | Intel Core i7-8650U CPU 1.90GHz (Kaby Lake R) | @jeffhandley | | Windows 10.0.19041.450 | X64 | Intel Core i7-8700 CPU 3.20GHz (Coffee Lake) | @jeffhandley | | ubuntu 18.04 | X64 | Intel Xeon CPU E5-1650 v4 3.60GHz | @adamsitnik | | manjaro | X64 | Intel Core i7-4771 CPU 3.50GHz (Haswell) | @ManickaP | | pop 20.04 | X64 | Intel Core i7-6600U CPU 2.60GHz (Skylake) | @carlossanlop | | alpine 3.11 (WSL2) | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | @danmosemsft | | ubuntu 18.04 (WSL2) | X64 | Intel Core i7-7700 CPU 3.60GHz (Kaby Lake) | @danmosemsft | | ubuntu 16.04 | Arm64 | Qualcomm Centriq | @adamsitnik | | ubuntu 18.04 (WSL2) | Arm64 | Microsoft SQ1 3.0 GHz (Surface Pro X) | @carlossanlop | | ubuntu 20.04 (WSL2) | Arm64 | Microsoft SQ1 3.0 GHz (Surface Pro X) | @pgovind | | Windows 10.0.18363.959 | X86 | Intel Xeon CPU E5-1650 v4 3.60GHz | @adamsitnik | | Windows 10.0.19041.450 | X86 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | @adamsitnik | | Windows 10.0.18363.1016 | Arm | Microsoft SQ1 3.0 GHz (Surface Pro X) | @adamsitnik | | macOS Catalina 10.15.6 | X64 | Intel Core i5-4278U CPU 2.60GHz (Haswell) | @jeffhandley | | macOS Catalina 10.15.6 | X64 | Intel Core i7-4870HQ CPU 2.50GHz (Haswell) | @carlossanlop | | macOS Mojave 10.14.5 | X64 | Intel Core i7-5557U CPU 3.10GHz (Broadwell) | @adamsitnik | Everyone interested can download the data from [here](https://microsofteur-my.sharepoint.com/:u:/g/personal/adsitnik_microsoft_com/Ea0pIss74N9GsDywOk3Wq_0BFcyFSEJ6m5xFMseha1T0IA?e=bWQpIp). The full report generated by the tool is available [here](https://microsofteur-my.sharepoint.com/:t:/g/personal/adsitnik_microsoft_com/EY3O7vVrCX9OkulTAowPbToBGFmQrTfosY0LDhXBxkytAQ?e=Aj0BRS). Moreover, **the full historical data turned out to be extremely useful**. I've used it every time I was not sure whether something was a regression or just unstable|multimodal benchmark: * [Windows 10 x64](https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/master_x64_Windows%2010.0.18362/PerfLabTests.BlockCopyPerf.CallBlockCopy(numElements:%201000).html) * [Ubuntu 18.04 x64](https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/09_01_2020/refs/heads/master_x64_ubuntu%2018.04/System.Threading.Tests.Perf_CancellationToken.html) * [Ubuntu 18.04 ARM64](https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/master_arm64_ubuntu%2018.04/AllTestindex.html) (added yesterday) * [Windows 10 x86](https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/master_x86_Windows%2010.0.18362/PerfLabTests.BlockCopyPerf.CallBlockCopy(numElements:%201000).html) # Regressions ## Already fixed - [x] `System.Collections.Contains*`, `System.Memory.SequenceReader.TryReadTo`, `System.Text.Json.Tests.Perf_Segment.ReadSingleSegmentSequenceByN` * was a 32 bit issue only (both `x86` and `ARM`) * detected by the bot, reported in https://github.com/DrewScoggins/performance-2/issues/910#issuecomment-676413559 * confirmed: https://github.com/DrewScoggins/performance-2/issues/910#issuecomment-677854221 * transffered to runtime repo: https://github.com/dotnet/runtime/issues/41167 * fixed in https://github.com/dotnet/runtime/pull/41198 * backported to 5.0 in https://github.com/dotnet/runtime/pull/41254 - [x] `System.Collections.CtorGivenSize<Int32>.Array(Size: 512)` * specific to Apline only * created an issue https://github.com/dotnet/runtime/issues/41398 * confirmed by @jkotas to be not WSL specific, but a much bigger Alpine perf problem * it has shown that an increased number of Gen 0 collections is a valuable metric to detect regressions * fixed in https://github.com/dotnet/runtime/pull/41532 * backported to 5.0 https://github.com/dotnet/runtime/pull/41547 * created https://github.com/dotnet/runtime/issues/41708 to add unit tests that ensure that this problem is not coming back - [x] `System.Numerics.Tests.Perf_Quaternion.Conjugate` and `System.Numerics.Tests.Perf_Quaternion.Negat*` * not reported by the bot because it's a brand new benchmark and we did not have historical data at the time of my investigation * issue created: https://github.com/dotnet/runtime/issues/41738 * fixed in https://github.com/dotnet/runtime/pull/41829 * backported to 5.0-rc2 in https://github.com/dotnet/runtime/pull/41885 - [x] `Directory.EnumerateFiles` * not reported by the bot, most probably because it was a very fresh regression * issue created: https://github.com/dotnet/runtime/issues/41739 * fixed in https://github.com/dotnet/runtime/issues/41739 * backported to 5.0-rc2 in https://github.com/dotnet/runtime/pull/41820 - [x] `ByteMark.BenchIDEAEncryption` * not reported by the bot, most probably because it was a very fresh regression * issue created: https://github.com/dotnet/runtime/issues/41677 * fixed in https://github.com/dotnet/runtime/pull/40871 * backported to 5.0-rc2 in https://github.com/dotnet/runtime/pull/41838 - [x] `System.Text.Perf_Utf8Encoding` * not detected by the bot because it was not enabled for ARM yet * issue created: https://github.com/dotnet/runtime/issues/41699 * fixed in https://github.com/dotnet/runtime/pull/42052 * backported to 5.0-rc2 in #42064 ## Investigation in progress - [ ] `System.Text.Encodings.Web.Tests.Perf_Encoders.EncodeUtf8` * not reported by the bot because it was a brand new benchmark and we did not have historical data at the time of my investigation * issue created: https://github.com/dotnet/runtime/issues/41104 * investigation is in progress - [ ] `System.Memory.Slice` * not detected by the bot because it was not enabled for ARM yet * seems to be ARM64-specific, created an issue https://github.com/dotnet/runtime/issues/41704 * investigation is in progress - [ ] `PerfLabTests.CastingPerf2.CastingPerf.IntObj` * not detected by the bot because it was not enabled for ARM yet * seems to be ARM64-specific, created an issue https://github.com/dotnet/runtime/issues/41706 * investigation is in progress ## By design or Acceptable - [ ] ICU-related regressions * `System.Globalization.Tests.StringSearch`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/37819 * `System.Memory.ReadOnlySpan.IndexOfString`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/39724 * `System.Globalization.Tests.Perf_DateTimeCultureInfo.Parse(culturestring: ja)`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/37807 * `System.Globalization.Tests.StringEquality`: detected by the bot, reported in https://github.com/dotnet/runtime/issues/39038 * I've created one uber issue to track all of them in one place: https://github.com/dotnet/runtime/issues/40942 * `OrdinalIgnoreCase` has been optimized in https://github.com/dotnet/runtime/pull/40962 * TODO: doc update still required - [x] `System.Linq.Tests.Perf_Enumerable.FirstWithPredicate_LastElementMatches(input: IOrderedEnumerable)` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39032 * closed, by design: removed the `O(N log N)` cost of the `OrderBy` https://github.com/dotnet/runtime/issues/39032#issuecomment-656678750 - [x] `System.Collections.Tests.Perf_BitArray.*(Size: 4)` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37813 * closed, by design: introduction of vectorization has increased the cost of operations for small inputs: https://github.com/dotnet/runtime/issues/37813#issuecomment-656370853 - [x] `System.Threading.Tests.Perf_Thread.GetCurrentProcessorId` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37804 * closed, by design: precision was improved at a cost of acceptable minor perf regression: https://github.com/dotnet/runtime/issues/37804#issuecomment-643448336 - [x] `PerfLabTests.CastingPerf.CheckIsInstAnyIsInterfaceNo`, `PerfLabTests.CastingPerf.CheckObjIsInterfaceNo` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37803 * closed, by design: known tradeoff: https://github.com/dotnet/runtime/issues/37803#issuecomment-670209689 - [x] `System.Net.NetworkInformation.Tests.PhysicalAddressTests.PAShort` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39720 * closed, acceptable for improved code reuse https://github.com/dotnet/runtime/issues/39720#issuecomment-671535578 * benchmark for 1 byte removed, added 6 bytes in https://github.com/dotnet/performance/pull/1490 - [x] `System.Numerics.Tests.Perf_Vector*.GetHashCodeBenchmark` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39035 and https://github.com/dotnet/runtime/issues/39029 * closed, "it should not be used" https://github.com/dotnet/runtime/issues/39029#issuecomment-656408464 - [ ] `System.Net.Primitives.Tests.CredentialCacheTests.ForEach(uriCount: 0, hostPortCount: 0)` * detected by the bot, reported in https://github.com/DrewScoggins/performance-2/issues/510 * confirmed: https://github.com/DrewScoggins/performance-2/issues/510#issuecomment-680783031 * awaiting the transfer to runtime repo. Most probably a by-design regression. ## Moved to 6.0 - [x] `System.Tests.Perf_Char.GetUnicodeCategory(c: '?')` * detected and reported by the bot in https://github.com/DrewScoggins/performance-2/issues/574, I've created https://github.com/dotnet/runtime/issues/41107 * minor regression for non-ascii characters, moved to 6.0 - [x] `PerfLabTests.StackWalk.Walk` * detected by the bot and reported in https://github.com/dotnet/runtime/issues/39115 * confirmed in https://github.com/dotnet/runtime/issues/39115#issuecomment-677857078 * specific to everything that is not Windows x64, rather not critical -> moved to 6.0: https://github.com/dotnet/runtime/issues/39115#issuecomment-682684126 - [x] `System.Tests.Perf_String.Replace_Char(text: "Hello", oldChar: 'l', newChar: '!')` * reported in https://github.com/dotnet/runtime/issues/37816 * confirmed in https://github.com/dotnet/runtime/issues/37816#issuecomment-680706314 * moved to 6.0 - [x] `System.Text.Perf_Utf8String.IsAscii(Input: EnglishAllAscii)` * not reported by the bot because it was a brand new benchmark and we did not have historical data at the time of my investigation * issue created: https://github.com/dotnet/runtime/issues/41388 * moved to 6.0 as `Utf8String` is still only experimental ## Unstable or multimodal benchmarks There was of course more of them, here are the ones that I've noted to use as Contract Tests in the near future (to reduce the noise produced by the bot): * `System.Buffers.Tests.RentReturnArrayPoolTests<Byte>.ProducerConsumer` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/39031 * asked for historical data to verify if it's multimodal or not https://github.com/dotnet/runtime/issues/39031#issuecomment-680712003 * thanks to historical data provided it was possible to tell that it's unstable for x64 and bimodal for x86: https://github.com/dotnet/runtime/issues/39031#issuecomment-682207270 * `System.Memory.ReadOnlySequence.Slice_Repeat_StartPosition_And_EndPosition(Segment: Multiple)` * quite unstable benchmark, I've verified that 5.0 codegen is better * `PerfLabTests.BlockCopyPerf.CallBlockCopy` * detected by the bot, reported in https://github.com/dotnet/runtime/issues/37808 * copying 0 elements does not add value: https://github.com/dotnet/runtime/issues/37808#issuecomment-654424436 * test case for copying 0 elements removed in https://github.com/dotnet/performance/pull/1465 * closed as unstable based on full historical data: https://github.com/dotnet/runtime/issues/37808#issuecomment-685014522 * `System.Tests.Perf_String.Trim_CharArr(s: "Test", c: [' ', ' '])` * multimodal benchmark, needs a rewrite as stated long time ago: https://github.com/dotnet/runtime/issues/13135 * `System.Threading.Tests.Perf_Interlocked.CompareExchange_long` * the benchmark typically reports `10ns`, but sometimes `x100` that. Only for `x86`. I need logs to verify whether it's a BDN bug or not. * issue created https://github.com/dotnet/performance/issues/1497 * `System.Memory.Span<Int32>.IndexOfValue(Size: 512)` * reported in https://github.com/dotnet/runtime/issues/39722 * confirmed that it was due to code alignment change in https://github.com/dotnet/runtime/issues/39722#issuecomment-674999435 * `Benchstone.BenchI.Fib.Test` * perfectly multimodal, great example for a contract test # Summary * The bot has reported all major performance issues for the configurations that it was enabled for (Windows x64, x86, and Ubuntu x64). Great work @DrewScoggins! * The full historical data turned out to be extremely useful to exclude all false positives for multimodal and unstable benchmarks. * We have missed one important x86 bug during triaging (human error), but it got discovered during the study (https://github.com/dotnet/runtime/issues/41167#issuecomment-679285756). To avoid such problems in the future and to enable the bot in the runtime repo, the noise of the bot needs to be reduced. Currently, it's quite high, mostly due to the multimodal nature of the benchmarks. * The study has detected relatively many new ARM64 perf problems at a late stage of the release. The sooner we enable the bot for ARM64, the better. Moreover, we should be more frequently asking for ARM64 results when reviewing big changes that affect the performance of frequently used features (like sorting the arrays). * The study has shown that measuring the performance of `GNU libc` based Linux distros like Ubuntu is not enough to detect `musl libc` specific regressions. We should consider adding Alpine runs to the perf lab. * This time no important issues specific to `macOS` and different CPU families were discovered. It has proven that the perf lab has good hardware coverage. * The Alpine regression has shown that an increased number of Gen 0 collections can be a very valuable metric to detect regressions. We should consider extending the bot to use it. **Big thanks to everyone involved!**
non_priority
net microbenchmarks performance study report goals the main goal of my study was to ensure that we ship net without any performance regressions and validate whether in the near future we can fully rely on the regressions auto filing bot written by drewscoggins my other goal was to get net library team members involved and keep on growing the performance culture tl dr the bot is doing a great job in detecting regressions most serious regressions have been already fixed however a few investigations are still in progress methodology and how it evolved in i had the pleasure to review andreyakinshin book the statistics for performance engineers and performance analysis and performance testing chapters inspired me to a small tool called the tool uses the statistical test to detect performance regressions in results exported by benchmarkdotnet it s being used or at least it should as part of our to prevent introducing regressions to net in i was asked by danmosemsft to verify net core performance initially i’ve run all the using a single machine with dual boot for windows and ubuntu and used the results comparer to find regressions it very quickly turned out that such a sample was way too small to make sure that we don’t have any regressions some benchmarks were simply unstable some architectures like arm and were simply not covered other linux distros and cpu families were also not covered then i’ve run the benchmarks on all the pcs laptops and vms that i could access but i was still missing amd and arm results so i ve asked tannergooding and bruceforstall for help tannergooding has run the benchmarks on all his amd machines bruceforstall has provided me access to a document that explains how to use arm machines owned by the jit team this turned out to be an invaluable help as i ve used these machines many many times including this year during the investigation after having enough samples to cover our matrix of supported oses and architectures i’ve built a simple console app on top of resultscomparer source code available the tool uses the very same statistical test to detect regressions aggregates the results from all different configurations and sorts them from the biggest regression to the biggest improvement such approach allows for very quick identification of regressions of all kinds affecting every configuration system linq tests perf enumerable firstwithpredicate lastelementmatches input iorderedenumerable result base diff ratio operating system bit slower windows slower windows slower windows slower windows slower windows slower windows slower windows slower windows slower ubuntu slower manjaro slower pop slower alpine slower ubuntu slower ubuntu slower ubuntu slower ubuntu slower ubuntu slower ubuntu slower windows slower windows slower windows arm slower macos catalina slower macos catalina slower macos mojave affecting specific os families windows unix system globalization tests stringsearch isprefix differentfirstchar options en us ignoresymbols false result base diff ratio operating system bit slower windows slower windows slower windows slower windows slower windows slower windows slower windows slower windows same ubuntu same manjaro same pop same alpine same ubuntu same ubuntu same ubuntu same ubuntu same ubuntu faster ubuntu slower windows slower windows slower windows arm same macos catalina same macos catalina same macos mojave affecting specific linux distros system threading tests perf cancellationtoken cancel result base diff ratio operating system bit same windows same windows same windows same windows same windows same windows same windows same windows same ubuntu faster manjaro same pop slower alpine same ubuntu same ubuntu same ubuntu same ubuntu same ubuntu faster ubuntu same windows same windows same windows arm same macos catalina same macos catalina same macos mojave affecting specific cpu families system buffers text tests numberofbytes result base diff ratio operating system bit processor name same windows amd ryzen same windows intel xeon cpu same windows intel xeon cpu same windows intel core cpu broadwell slower windows intel core cpu skylake same windows intel core cpu kaby lake same windows intel core cpu kaby lake r slower windows intel core cpu coffee lake same ubuntu intel xeon cpu same manjaro intel core cpu haswell slower pop intel core cpu skylake same alpine intel core cpu kaby lake same ubuntu intel core cpu kaby lake same ubuntu unknown processor same ubuntu unknown processor same ubuntu unknown processor same ubuntu unknown processor same ubuntu unknown processor same windows intel xeon cpu same windows intel core cpu broadwell same windows arm microsoft ghz same macos catalina intel core cpu haswell same macos catalina intel core cpu haswell same macos mojave intel core cpu broadwell using the tool had one major flaw it was not automated and hence we were finding out about the regressions only when we searched for them this has been recognized and a new project has been started in drewscoggins started implementing a github bot that would be using the data gathered from performance lab a set of machines owned by net performance team microbenchmark runs to detect and auto file the regressions so far the bot was reporting new issues in a and once a week the workgroup led by drewscoggins that consisted of andyayersms kunalspathak tannergooding any myself was going through the list and triaging the issues issues that were seemed as actual regressions were labeled as and were later moved by drewscoggins to the a few weeks ago we were getting close to code freeze for net and i have asked myself a question are we sure that the bot has reported all possible regressions for the bot is using different statistical methods to detect regressions and so far it has been enabled only for windows ubuntu and windows so i ve decided to spend some time and use the old tool that i wrote to verify it to increase the sample size and get other net libraries team members involved i ve simply asked the team to run the benchmarks and share the results with me running the performance repo microbenchmarks against the latest net core sdk is super easy thanks to a python script by jorive the script downloads the right sdk and starts benchmarking with cleared environment variables cmd git clone performance scripts benchmarks ci py f filter data the data i ve received from the net libraries team members allowed me a big part of the entire operating system arch processor name provided by windows amd ryzen tannergooding windows intel xeon cpu adamsitnik windows intel core cpu broadwell adamsitnik windows intel core cpu skylake grabyourpitchforks windows intel core cpu kaby lake danmosemsft windows intel core cpu kaby lake r jeffhandley windows intel core cpu coffee lake jeffhandley ubuntu intel xeon cpu adamsitnik manjaro intel core cpu haswell manickap pop intel core cpu skylake carlossanlop alpine intel core cpu kaby lake danmosemsft ubuntu intel core cpu kaby lake danmosemsft ubuntu qualcomm centriq adamsitnik ubuntu microsoft ghz surface pro x carlossanlop ubuntu microsoft ghz surface pro x pgovind windows intel xeon cpu adamsitnik windows intel core cpu broadwell adamsitnik windows arm microsoft ghz surface pro x adamsitnik macos catalina intel core cpu haswell jeffhandley macos catalina intel core cpu haswell carlossanlop macos mojave intel core cpu broadwell adamsitnik everyone interested can download the data from the full report generated by the tool is available moreover the full historical data turned out to be extremely useful i ve used it every time i was not sure whether something was a regression or just unstable multimodal benchmark added yesterday regressions already fixed system collections contains system memory sequencereader tryreadto system text json tests perf segment readsinglesegmentsequencebyn was a bit issue only both and arm detected by the bot reported in confirmed transffered to runtime repo fixed in backported to in system collections ctorgivensize array size specific to apline only created an issue confirmed by jkotas to be not wsl specific but a much bigger alpine perf problem it has shown that an increased number of gen collections is a valuable metric to detect regressions fixed in backported to created to add unit tests that ensure that this problem is not coming back system numerics tests perf quaternion conjugate and system numerics tests perf quaternion negat not reported by the bot because it s a brand new benchmark and we did not have historical data at the time of my investigation issue created fixed in backported to in directory enumeratefiles not reported by the bot most probably because it was a very fresh regression issue created fixed in backported to in bytemark benchideaencryption not reported by the bot most probably because it was a very fresh regression issue created fixed in backported to in system text perf not detected by the bot because it was not enabled for arm yet issue created fixed in backported to in investigation in progress system text encodings web tests perf encoders not reported by the bot because it was a brand new benchmark and we did not have historical data at the time of my investigation issue created investigation is in progress system memory slice not detected by the bot because it was not enabled for arm yet seems to be specific created an issue investigation is in progress perflabtests castingperf intobj not detected by the bot because it was not enabled for arm yet seems to be specific created an issue investigation is in progress by design or acceptable icu related regressions system globalization tests stringsearch detected by the bot reported in system memory readonlyspan indexofstring detected by the bot reported in system globalization tests perf datetimecultureinfo parse culturestring ja detected by the bot reported in system globalization tests stringequality detected by the bot reported in i ve created one uber issue to track all of them in one place ordinalignorecase has been optimized in todo doc update still required system linq tests perf enumerable firstwithpredicate lastelementmatches input iorderedenumerable detected by the bot reported in closed by design removed the o n log n cost of the orderby system collections tests perf bitarray size detected by the bot reported in closed by design introduction of vectorization has increased the cost of operations for small inputs system threading tests perf thread getcurrentprocessorid detected by the bot reported in closed by design precision was improved at a cost of acceptable minor perf regression perflabtests castingperf checkisinstanyisinterfaceno perflabtests castingperf checkobjisinterfaceno detected by the bot reported in closed by design known tradeoff system net networkinformation tests physicaladdresstests pashort detected by the bot reported in closed acceptable for improved code reuse benchmark for byte removed added bytes in system numerics tests perf vector gethashcodebenchmark detected by the bot reported in and closed it should not be used system net primitives tests credentialcachetests foreach uricount hostportcount detected by the bot reported in confirmed awaiting the transfer to runtime repo most probably a by design regression moved to system tests perf char getunicodecategory c detected and reported by the bot in i ve created minor regression for non ascii characters moved to perflabtests stackwalk walk detected by the bot and reported in confirmed in specific to everything that is not windows rather not critical moved to system tests perf string replace char text hello oldchar l newchar reported in confirmed in moved to system text perf isascii input englishallascii not reported by the bot because it was a brand new benchmark and we did not have historical data at the time of my investigation issue created moved to as is still only experimental unstable or multimodal benchmarks there was of course more of them here are the ones that i ve noted to use as contract tests in the near future to reduce the noise produced by the bot system buffers tests rentreturnarraypooltests producerconsumer detected by the bot reported in asked for historical data to verify if it s multimodal or not thanks to historical data provided it was possible to tell that it s unstable for and bimodal for system memory readonlysequence slice repeat startposition and endposition segment multiple quite unstable benchmark i ve verified that codegen is better perflabtests blockcopyperf callblockcopy detected by the bot reported in copying elements does not add value test case for copying elements removed in closed as unstable based on full historical data system tests perf string trim chararr s test c multimodal benchmark needs a rewrite as stated long time ago system threading tests perf interlocked compareexchange long the benchmark typically reports but sometimes that only for i need logs to verify whether it s a bdn bug or not issue created system memory span indexofvalue size reported in confirmed that it was due to code alignment change in benchstone benchi fib test perfectly multimodal great example for a contract test summary the bot has reported all major performance issues for the configurations that it was enabled for windows and ubuntu great work drewscoggins the full historical data turned out to be extremely useful to exclude all false positives for multimodal and unstable benchmarks we have missed one important bug during triaging human error but it got discovered during the study to avoid such problems in the future and to enable the bot in the runtime repo the noise of the bot needs to be reduced currently it s quite high mostly due to the multimodal nature of the benchmarks the study has detected relatively many new perf problems at a late stage of the release the sooner we enable the bot for the better moreover we should be more frequently asking for results when reviewing big changes that affect the performance of frequently used features like sorting the arrays the study has shown that measuring the performance of gnu libc based linux distros like ubuntu is not enough to detect musl libc specific regressions we should consider adding alpine runs to the perf lab this time no important issues specific to macos and different cpu families were discovered it has proven that the perf lab has good hardware coverage the alpine regression has shown that an increased number of gen collections can be a very valuable metric to detect regressions we should consider extending the bot to use it big thanks to everyone involved
0
194,995
14,698,531,033
IssuesEvent
2021-01-04 06:40:32
isontheline/pro.webssh.net
https://api.github.com/repos/isontheline/pro.webssh.net
closed
Fn keys : move access to a better place
enhancement testflight-fixed
**Describe the feature** Fn keys on the navigation bar is not a good idea :/ Instead add them to : * the virtual keyboard OR * the top right floating menu OR * the contextual menu
1.0
Fn keys : move access to a better place - **Describe the feature** Fn keys on the navigation bar is not a good idea :/ Instead add them to : * the virtual keyboard OR * the top right floating menu OR * the contextual menu
non_priority
fn keys move access to a better place describe the feature fn keys on the navigation bar is not a good idea instead add them to the virtual keyboard or the top right floating menu or the contextual menu
0
212,210
7,229,359,057
IssuesEvent
2018-02-11 19:04:23
Team-2502/RobotCode2018
https://api.github.com/repos/Team-2502/RobotCode2018
closed
PIDs are strange
PID bug calibration / tuning high priority ¯\_(ツ)_/¯
When using PID for tuning wheel velocity, the velocities always oscillate at a value too small. For example, at the target speed of 1 RPM, the wheels actually go at ~.75 RPM (this proportion is not constant at different speeds).
1.0
PIDs are strange - When using PID for tuning wheel velocity, the velocities always oscillate at a value too small. For example, at the target speed of 1 RPM, the wheels actually go at ~.75 RPM (this proportion is not constant at different speeds).
priority
pids are strange when using pid for tuning wheel velocity the velocities always oscillate at a value too small for example at the target speed of rpm the wheels actually go at rpm this proportion is not constant at different speeds
1
21,711
6,208,822,397
IssuesEvent
2017-07-07 01:23:08
ahmedahamid/test
https://api.github.com/repos/ahmedahamid/test
closed
Cannot connect to TFS server anymore
bug CodePlexMigrationInitiated impact: High
Since you guys changed the user names to automatically lower-case the first letter (on the CodePlex site itself), I cannot connect to the TFS01. I have tried IDisposable (what I signed up as--which used to work), iDisposable (what is show as I login--JavaScript client changing this?), and idisposable (just a guess). I think something is broken. #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '338' Vote count: '0'
1.0
Cannot connect to TFS server anymore - Since you guys changed the user names to automatically lower-case the first letter (on the CodePlex site itself), I cannot connect to the TFS01. I have tried IDisposable (what I signed up as--which used to work), iDisposable (what is show as I login--JavaScript client changing this?), and idisposable (just a guess). I think something is broken. #### Migrated CodePlex Work Item Details CodePlex Work Item ID: '338' Vote count: '0'
non_priority
cannot connect to tfs server anymore since you guys changed the user names to automatically lower case the first letter on the codeplex site itself i cannot connect to the i have tried idisposable what i signed up as which used to work idisposable what is show as i login javascript client changing this and idisposable just a guess i think something is broken migrated codeplex work item details codeplex work item id vote count
0
536,012
15,702,653,712
IssuesEvent
2021-03-26 12:55:09
HabitRPG/habitica-ios
https://api.github.com/repos/HabitRPG/habitica-ios
closed
Reminders go off on the day they're set up, not on the day scheduled
Help wanted Priority: medium Type: Bug
User report: > When I set up an item in my todo list for a future date and set a reminder, the reminder goes off on the day I set up the todo instead of the date the item is scheduled. > > > The following lines help us find and squash the Bug you encountered. Please do not delete/change them. > iOS Version: 14.3 > Device: iPhone13,3 > App Version: 3.1.3 (532) > User ID: 2894d56a-74c7-4d89-bed1-c0c1dcddf898 > Level: 95 > Class: rogue > Is in Inn: false > Uses Costume: true > Custom Day Start: 0
1.0
Reminders go off on the day they're set up, not on the day scheduled - User report: > When I set up an item in my todo list for a future date and set a reminder, the reminder goes off on the day I set up the todo instead of the date the item is scheduled. > > > The following lines help us find and squash the Bug you encountered. Please do not delete/change them. > iOS Version: 14.3 > Device: iPhone13,3 > App Version: 3.1.3 (532) > User ID: 2894d56a-74c7-4d89-bed1-c0c1dcddf898 > Level: 95 > Class: rogue > Is in Inn: false > Uses Costume: true > Custom Day Start: 0
priority
reminders go off on the day they re set up not on the day scheduled user report when i set up an item in my todo list for a future date and set a reminder the reminder goes off on the day i set up the todo instead of the date the item is scheduled the following lines help us find and squash the bug you encountered please do not delete change them ios version device app version user id level class rogue is in inn false uses costume true custom day start
1
719,136
24,748,065,155
IssuesEvent
2022-10-21 11:27:25
ooni/explorer
https://api.github.com/repos/ooni/explorer
closed
Search queries should be done client side
enhancement priority/medium
Currently when I access https://explorer.ooni.org/search, the queries are being performed server-side, instead these should be done client-side so that we don't block on rendering some content to the user. This may also help solve the fact that while the search page is rendering it is not possible to toggle the search parameters.
1.0
Search queries should be done client side - Currently when I access https://explorer.ooni.org/search, the queries are being performed server-side, instead these should be done client-side so that we don't block on rendering some content to the user. This may also help solve the fact that while the search page is rendering it is not possible to toggle the search parameters.
priority
search queries should be done client side currently when i access the queries are being performed server side instead these should be done client side so that we don t block on rendering some content to the user this may also help solve the fact that while the search page is rendering it is not possible to toggle the search parameters
1
813,043
30,443,425,307
IssuesEvent
2023-07-15 11:04:45
Counselllor/Counsellor-Web
https://api.github.com/repos/Counselllor/Counsellor-Web
closed
I will be creating landing page
gssoc23 assigned level3 priority: high
## Issue I will be creating the first page of the website ## Screenshot ![218539171-2469465b-d840-4c10-99e9-df3af37c49d4.png](https://github.com/Counselllor/Counsellor-Web/assets/85278243/f9628c9c-ffad-4d4e-babb-a7e514cbd608) Please assign me this task so that I can create the above mentioned page I am in GSSOC'23
1.0
I will be creating landing page - ## Issue I will be creating the first page of the website ## Screenshot ![218539171-2469465b-d840-4c10-99e9-df3af37c49d4.png](https://github.com/Counselllor/Counsellor-Web/assets/85278243/f9628c9c-ffad-4d4e-babb-a7e514cbd608) Please assign me this task so that I can create the above mentioned page I am in GSSOC'23
priority
i will be creating landing page issue i will be creating the first page of the website screenshot please assign me this task so that i can create the above mentioned page i am in gssoc
1
418,313
12,196,187,681
IssuesEvent
2020-04-29 18:38:56
googleapis/google-cloud-ruby
https://api.github.com/repos/googleapis/google-cloud-ruby
closed
NoMethodError (undefined method `shutdown?' for nil:NilClass) after stopping async_writer
api: logging priority: p2 type: bug
#### Environment details - OS: ubuntu - Ruby version: 2.6.3 - Gem name and version: google-cloud-logging #### Steps to reproduce 1. setup `async_writer` normally. 2. without logging anything, run `.stop!` #### Code example ```ruby logging = Google::Cloud::Logging.new #project id async = logging.async_writer async.stop! ``` #### Full backtrace ```sh irb(main):003:0> async.stop! Traceback (most recent call last): 1: from (irb):3 NoMethodError (undefined method `shutdown?' for nil:NilClass) ``` #### Note This doesn't happen if something had been logged
1.0
NoMethodError (undefined method `shutdown?' for nil:NilClass) after stopping async_writer - #### Environment details - OS: ubuntu - Ruby version: 2.6.3 - Gem name and version: google-cloud-logging #### Steps to reproduce 1. setup `async_writer` normally. 2. without logging anything, run `.stop!` #### Code example ```ruby logging = Google::Cloud::Logging.new #project id async = logging.async_writer async.stop! ``` #### Full backtrace ```sh irb(main):003:0> async.stop! Traceback (most recent call last): 1: from (irb):3 NoMethodError (undefined method `shutdown?' for nil:NilClass) ``` #### Note This doesn't happen if something had been logged
priority
nomethoderror undefined method shutdown for nil nilclass after stopping async writer environment details os ubuntu ruby version gem name and version google cloud logging steps to reproduce setup async writer normally without logging anything run stop code example ruby logging google cloud logging new project id async logging async writer async stop full backtrace sh irb main async stop traceback most recent call last from irb nomethoderror undefined method shutdown for nil nilclass note this doesn t happen if something had been logged
1
370,361
25,902,444,450
IssuesEvent
2022-12-15 07:21:40
VisualGameData/VIGAD
https://api.github.com/repos/VisualGameData/VIGAD
closed
Update SRS accordingly
documentation Phase: Construction RUP: Project Management
# Issue description *Describe your issue in detail here* Our SRS should be up-to-date and be clearer, therefore we should be update ours. Maybe get inspired here from : https://github.com/nilskre/CommonPlayground/blob/dev/docs/SoftwareRequirementsSpecification.md How they did it. So please update it accordingly # Definition Of Ready (DoR) *This issue can be worked on if* - [x] a developer is assigned - [x] the issue is labeled - [x] the issue is assigned to the current sprint - [x] the issue is estimated - [x] a meaningful title has been set - [x] the issue is described in detail
1.0
Update SRS accordingly - # Issue description *Describe your issue in detail here* Our SRS should be up-to-date and be clearer, therefore we should be update ours. Maybe get inspired here from : https://github.com/nilskre/CommonPlayground/blob/dev/docs/SoftwareRequirementsSpecification.md How they did it. So please update it accordingly # Definition Of Ready (DoR) *This issue can be worked on if* - [x] a developer is assigned - [x] the issue is labeled - [x] the issue is assigned to the current sprint - [x] the issue is estimated - [x] a meaningful title has been set - [x] the issue is described in detail
non_priority
update srs accordingly issue description describe your issue in detail here our srs should be up to date and be clearer therefore we should be update ours maybe get inspired here from how they did it so please update it accordingly definition of ready dor this issue can be worked on if a developer is assigned the issue is labeled the issue is assigned to the current sprint the issue is estimated a meaningful title has been set the issue is described in detail
0
21,816
3,756,895,313
IssuesEvent
2016-03-13 17:08:47
gustafl/lexeme
https://api.github.com/repos/gustafl/lexeme
opened
Consider making lexical categories language-dependant
design future
I just learned that Japanese don't have prepositions but _postpositions_. In the very long run, we may have to accept that the nine lexical categories are not the same for every language out there.
1.0
Consider making lexical categories language-dependant - I just learned that Japanese don't have prepositions but _postpositions_. In the very long run, we may have to accept that the nine lexical categories are not the same for every language out there.
non_priority
consider making lexical categories language dependant i just learned that japanese don t have prepositions but postpositions in the very long run we may have to accept that the nine lexical categories are not the same for every language out there
0
277,665
8,631,057,813
IssuesEvent
2018-11-22 05:52:08
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
closed
Reports incorrectly filtered if filter changes before search returns
Priority: 2 - Medium Status: 5 - Ready Type: Bug
If you use the form type filter in the Reports tab the results can include reports that should not be there. To reproduce: - Select an item from the form filter, wait for the results to be updated in the LHS - Select a different item from the form filter **and deselect the item selected first before the results return** - See that the reports list includes reports for the selected and un-selected form types: ![image](https://user-images.githubusercontent.com/789512/37928342-8a0e67a8-310a-11e8-91b1-9e5d9bcd8d82.png) Tested on `beta.dev` running `2.15.0-beta.1`
1.0
Reports incorrectly filtered if filter changes before search returns - If you use the form type filter in the Reports tab the results can include reports that should not be there. To reproduce: - Select an item from the form filter, wait for the results to be updated in the LHS - Select a different item from the form filter **and deselect the item selected first before the results return** - See that the reports list includes reports for the selected and un-selected form types: ![image](https://user-images.githubusercontent.com/789512/37928342-8a0e67a8-310a-11e8-91b1-9e5d9bcd8d82.png) Tested on `beta.dev` running `2.15.0-beta.1`
priority
reports incorrectly filtered if filter changes before search returns if you use the form type filter in the reports tab the results can include reports that should not be there to reproduce select an item from the form filter wait for the results to be updated in the lhs select a different item from the form filter and deselect the item selected first before the results return see that the reports list includes reports for the selected and un selected form types tested on beta dev running beta
1
531,294
15,444,338,957
IssuesEvent
2021-03-08 10:16:39
AY2021S2-CS2103T-T12-4/tp
https://api.github.com/repos/AY2021S2-CS2103T-T12-4/tp
closed
Bug: Response time adds multiple seconds
priority.High severity.High type.Bug
To recreate, open the close, do a few commands and close the app multiple times. Sample of error: { "endpoints" : [ { "method" : "GET", "address" : "jakjdfl", "data" : "somedata", "headers" : [ ], "tagged" : [ ], "response" : { "protocolVersion" : "", "statusCode" : "", "reasonPhrase" : "", "statusLine" : "", "responseEntity" : "", "responseTime" : " seconds seconds seconds" } }, { "method" : "GET", "address" : "testingurl", "data" : "", "headers" : [ ], "tagged" : [ ], "response" : { "protocolVersion" : "", "statusCode" : "", "reasonPhrase" : "", "statusLine" : "", "responseEntity" : "", "responseTime" : " seconds" } }, { "method" : "POST", "address" : "urlwdata", "data" : "testingdata", "headers" : [ ], "tagged" : [ ], "response" : { "protocolVersion" : "", "statusCode" : "", "reasonPhrase" : "", "statusLine" : "", "responseEntity" : "", "responseTime" : " seconds" } } ] }
1.0
Bug: Response time adds multiple seconds - To recreate, open the close, do a few commands and close the app multiple times. Sample of error: { "endpoints" : [ { "method" : "GET", "address" : "jakjdfl", "data" : "somedata", "headers" : [ ], "tagged" : [ ], "response" : { "protocolVersion" : "", "statusCode" : "", "reasonPhrase" : "", "statusLine" : "", "responseEntity" : "", "responseTime" : " seconds seconds seconds" } }, { "method" : "GET", "address" : "testingurl", "data" : "", "headers" : [ ], "tagged" : [ ], "response" : { "protocolVersion" : "", "statusCode" : "", "reasonPhrase" : "", "statusLine" : "", "responseEntity" : "", "responseTime" : " seconds" } }, { "method" : "POST", "address" : "urlwdata", "data" : "testingdata", "headers" : [ ], "tagged" : [ ], "response" : { "protocolVersion" : "", "statusCode" : "", "reasonPhrase" : "", "statusLine" : "", "responseEntity" : "", "responseTime" : " seconds" } } ] }
priority
bug response time adds multiple seconds to recreate open the close do a few commands and close the app multiple times sample of error endpoints method get address jakjdfl data somedata headers tagged response protocolversion statuscode reasonphrase statusline responseentity responsetime seconds seconds seconds method get address testingurl data headers tagged response protocolversion statuscode reasonphrase statusline responseentity responsetime seconds method post address urlwdata data testingdata headers tagged response protocolversion statuscode reasonphrase statusline responseentity responsetime seconds
1
546,696
16,017,844,926
IssuesEvent
2021-04-20 18:21:06
GoogleCloudPlatform/java-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/java-docs-samples
closed
com.example.trace.TraceSampleIT: testCreateAndRegisterFullSampled failed
:rotating_light: api: cloudtrace flakybot: issue priority: p1 samples type: bug
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 0059c8f41748088dda645ab19f738ac53f007f74 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/8d57b67d-cc48-4ea3-803d-8934047bf6c5), [Sponge](http://sponge2/8d57b67d-cc48-4ea3-803d-8934047bf6c5) status: failed <details><summary>Test output</summary><br><pre>java.lang.IllegalStateException: Stackdriver exporter is not registered. at com.example.trace.TraceSampleIT.tearDown(TraceSampleIT.java:42) </pre></details>
1.0
com.example.trace.TraceSampleIT: testCreateAndRegisterFullSampled failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 0059c8f41748088dda645ab19f738ac53f007f74 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/8d57b67d-cc48-4ea3-803d-8934047bf6c5), [Sponge](http://sponge2/8d57b67d-cc48-4ea3-803d-8934047bf6c5) status: failed <details><summary>Test output</summary><br><pre>java.lang.IllegalStateException: Stackdriver exporter is not registered. at com.example.trace.TraceSampleIT.tearDown(TraceSampleIT.java:42) </pre></details>
priority
com example trace tracesampleit testcreateandregisterfullsampled failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output java lang illegalstateexception stackdriver exporter is not registered at com example trace tracesampleit teardown tracesampleit java
1
404,428
11,857,328,929
IssuesEvent
2020-03-25 09:22:55
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.spectrum.net - site is not usable
browser-firefox-mobile engine-gecko form-v2-experiment ml-needsdiagnosis-false priority-normal
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/50669 --> <!-- @extra_labels: form-v2-experiment --> **URL**: https://www.spectrum.net/login **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Page will not load automatically <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/216c230d-e675-4a57-a74b-cd899115a3f2.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200316183117</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/3/b7c0216d-beb7-452e-900a-270ed30727f5) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.spectrum.net - site is not usable - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/50669 --> <!-- @extra_labels: form-v2-experiment --> **URL**: https://www.spectrum.net/login **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Page will not load automatically <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/216c230d-e675-4a57-a74b-cd899115a3f2.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200316183117</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/3/b7c0216d-beb7-452e-900a-270ed30727f5) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
site is not usable url browser version firefox mobile operating system android tested another browser yes other problem type site is not usable description page not loading correctly steps to reproduce page will not load automatically view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
805,179
29,510,327,902
IssuesEvent
2023-06-03 21:10:12
f1lem0n/falscify
https://api.github.com/repos/f1lem0n/falscify
closed
Write an about section for the website
feature priority
*Please read the following instructions carefully.* *Having detailed, yet not overcomplicated feedback will help us implement your idea efficiently.* 🚒 ***Before suggesting a new feature, please search for a similar feature requests in the [issues section](https://github.com/f1lem0n/falscify/issues).*** 👀️ ### Checklist ✅ - [x] I have read and consent to this project's [Code of Conduct](https://github.com/f1lem0n/falscify/blob/main/CODE_OF_CONDUCT.md). - [x] I added a very descriptive title. - [x] I searched through issues in this repository to find a similar feature request but did not find any. - [x] I searched through **falscify** documentation and found no such feature available. ### Feature description 📖 *Please describe your idea briefly and state why it is important to implement it.* Write an about section for the website containing: - mission - problem we are trying to solve - approach - team *Provide our team a list of changes/improvements we should take into consideration.* This change is outside of codebase and will be posted as a comment down below. ### Additional comments 💬 *You can provide some additional information in this section.* <br> *Thank you for your contribution!* 🤖
1.0
Write an about section for the website - *Please read the following instructions carefully.* *Having detailed, yet not overcomplicated feedback will help us implement your idea efficiently.* 🚒 ***Before suggesting a new feature, please search for a similar feature requests in the [issues section](https://github.com/f1lem0n/falscify/issues).*** 👀️ ### Checklist ✅ - [x] I have read and consent to this project's [Code of Conduct](https://github.com/f1lem0n/falscify/blob/main/CODE_OF_CONDUCT.md). - [x] I added a very descriptive title. - [x] I searched through issues in this repository to find a similar feature request but did not find any. - [x] I searched through **falscify** documentation and found no such feature available. ### Feature description 📖 *Please describe your idea briefly and state why it is important to implement it.* Write an about section for the website containing: - mission - problem we are trying to solve - approach - team *Provide our team a list of changes/improvements we should take into consideration.* This change is outside of codebase and will be posted as a comment down below. ### Additional comments 💬 *You can provide some additional information in this section.* <br> *Thank you for your contribution!* 🤖
priority
write an about section for the website please read the following instructions carefully having detailed yet not overcomplicated feedback will help us implement your idea efficiently 🚒 before suggesting a new feature please search for a similar feature requests in the 👀️ checklist ✅ i have read and consent to this project s i added a very descriptive title i searched through issues in this repository to find a similar feature request but did not find any i searched through falscify documentation and found no such feature available feature description 📖 please describe your idea briefly and state why it is important to implement it write an about section for the website containing mission problem we are trying to solve approach team provide our team a list of changes improvements we should take into consideration this change is outside of codebase and will be posted as a comment down below additional comments 💬 you can provide some additional information in this section thank you for your contribution 🤖
1
323,149
9,850,570,517
IssuesEvent
2019-06-19 08:29:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
mega.nz - site is not usable
browser-firefox engine-gecko priority-important
<!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://mega.nz/fm/undefined **Browser / Version**: Firefox 68.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: desktop inacsesible **Steps to Reproduce**: no se puede acceder a la nube para administrar carpetas [![Screenshot Description](https://webcompat.com/uploads/2019/6/be41ccfa-5f54-4ddf-997a-4399676a90ba-thumb.jpeg)](https://webcompat.com/uploads/2019/6/be41ccfa-5f54-4ddf-997a-4399676a90ba.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190613141208</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 318ms)" {file: "blob:https://mega.nz/a7489fbe-5a97-45bb-bc37-3a7618f4016e" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 285ms)" {file: "blob:https://mega.nz/8849462e-ae77-4f3d-97c9-2b222c4b2af9" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 138ms)" {file: "blob:https://mega.nz/8849462e-ae77-4f3d-97c9-2b222c4b2af9" line: 0}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 39915}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 9ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 132ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 3453ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 53ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[console.debug(This browser does not support advanced audio streaming..., IndexSizeError: Index or size is negative or greater than the allowed amount) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:36483:21]', u'[JavaScript Warning: "unreachable code after return statement" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 39915}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 13ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 22ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 163ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 172ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 84ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 85ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[console.debug(Download the React DevTools for a better development experience: https://fb.me/react-devtools) blob:https://mega.nz/3d1e992b-dabc-448e-8e0f-49f22c63a23a:6146:17]', u'[console.log(renderRecents:, undefined, undefined) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122194:13]', u'[console.time(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150140:13]', u'[console.timeEnd(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150168:25]', u'[console.time(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150270:21]', u'[console.timeEnd(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150306:21]', u'[console.time(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122326:17]', u'[console.timeEnd(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122334:17]', u'[console.error(Invalid node passed to M.buildtree) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:155635:17]', u'[console.error(Invalid node passed to M.buildtree) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:155635:17]', u'[console.log(renderRecents:, undefined, undefined) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122194:13]', u'[console.time(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150140:13]', u'[console.timeEnd(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150168:25]', u'[console.time(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150270:21]', u'[console.timeEnd(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150306:21]', u'[console.time(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122326:17]', u'[console.timeEnd(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122334:17]', u'[JavaScript Warning: "onmozfullscreenchange est obsoleto." {file: "https://mega.nz/fm/undefined" line: 0}]', u'[JavaScript Warning: "onmozfullscreenerror est obsoleto." {file: "https://mega.nz/fm/undefined" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
mega.nz - site is not usable - <!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://mega.nz/fm/undefined **Browser / Version**: Firefox 68.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: desktop inacsesible **Steps to Reproduce**: no se puede acceder a la nube para administrar carpetas [![Screenshot Description](https://webcompat.com/uploads/2019/6/be41ccfa-5f54-4ddf-997a-4399676a90ba-thumb.jpeg)](https://webcompat.com/uploads/2019/6/be41ccfa-5f54-4ddf-997a-4399676a90ba.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190613141208</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 318ms)" {file: "blob:https://mega.nz/a7489fbe-5a97-45bb-bc37-3a7618f4016e" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 285ms)" {file: "blob:https://mega.nz/8849462e-ae77-4f3d-97c9-2b222c4b2af9" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 138ms)" {file: "blob:https://mega.nz/8849462e-ae77-4f3d-97c9-2b222c4b2af9" line: 0}]', u'[JavaScript Warning: "unreachable code after return statement" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 39915}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 9ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 132ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 3453ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 53ms)" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 0}]', u'[console.debug(This browser does not support advanced audio streaming..., IndexSizeError: Index or size is negative or greater than the allowed amount) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:36483:21]', u'[JavaScript Warning: "unreachable code after return statement" {file: "blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a" line: 39915}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 13ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 22ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 163ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 172ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 84ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 85ms)" {file: "https://mega.nz/rsaasm.js" line: 0}]', u'[console.debug(Download the React DevTools for a better development experience: https://fb.me/react-devtools) blob:https://mega.nz/3d1e992b-dabc-448e-8e0f-49f22c63a23a:6146:17]', u'[console.log(renderRecents:, undefined, undefined) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122194:13]', u'[console.time(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150140:13]', u'[console.timeEnd(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150168:25]', u'[console.time(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150270:21]', u'[console.timeEnd(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150306:21]', u'[console.time(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122326:17]', u'[console.timeEnd(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122334:17]', u'[console.error(Invalid node passed to M.buildtree) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:155635:17]', u'[console.error(Invalid node passed to M.buildtree) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:155635:17]', u'[console.log(renderRecents:, undefined, undefined) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122194:13]', u'[console.time(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150140:13]', u'[console.timeEnd(recents:collectNodes) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150168:25]', u'[console.time(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150270:21]', u'[console.timeEnd(recents:collateActions) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:150306:21]', u'[console.time(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122326:17]', u'[console.timeEnd(recents:render) blob:https://mega.nz/b7987fa5-b081-42f0-ae6a-67f2fb849e6a:122334:17]', u'[JavaScript Warning: "onmozfullscreenchange est obsoleto." {file: "https://mega.nz/fm/undefined" line: 0}]', u'[JavaScript Warning: "onmozfullscreenerror est obsoleto." {file: "https://mega.nz/fm/undefined" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
mega nz site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description desktop inacsesible steps to reproduce no se puede acceder a la nube para administrar carpetas browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel aurora console messages u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u from with ❤️
1
360,456
10,693,056,987
IssuesEvent
2019-10-23 08:03:20
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
radikal.ru - see bug description
browser-focus-geckoview engine-gecko priority-normal
<!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.0; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://radikal.ru/video/3MJSVENbN7g **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android 7.0 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Secure Connection Failed **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
radikal.ru - see bug description - <!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.0; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://radikal.ru/video/3MJSVENbN7g **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android 7.0 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Secure Connection Failed **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
radikal ru see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description secure connection failed steps to reproduce browser configuration none from with ❤️
1
740,235
25,740,190,326
IssuesEvent
2022-12-08 05:15:24
googleapis/nodejs-ai-platform
https://api.github.com/repos/googleapis/nodejs-ai-platform
closed
AI platform create dataset image: "after all" hook: should delete the created dataset for "should create a new image dataset in the parent resource" failed
type: bug priority: p1 flakybot: issue api: vertex-ai
Note: #387 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 04f7c858217f1a3ce7b1072c7bf8946d39947532 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2c1b5aae-a46c-4a7a-a474-79da01e686eb), [Sponge](http://sponge2/2c1b5aae-a46c-4a7a-a474-79da01e686eb) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./delete-dataset.js undefined undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. Error: Command failed: node ./delete-dataset.js undefined undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/create-dataset-image.test.js:25:28) at Context.<anonymous> (test/create-dataset-image.test.js:50:5) at processImmediate (internal/timers.js:461:21)</pre></details>
1.0
AI platform create dataset image: "after all" hook: should delete the created dataset for "should create a new image dataset in the parent resource" failed - Note: #387 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 04f7c858217f1a3ce7b1072c7bf8946d39947532 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2c1b5aae-a46c-4a7a-a474-79da01e686eb), [Sponge](http://sponge2/2c1b5aae-a46c-4a7a-a474-79da01e686eb) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./delete-dataset.js undefined undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. Error: Command failed: node ./delete-dataset.js undefined undefined undefined 7 PERMISSION_DENIED: Permission denied: Consumer 'project:undefined' has been suspended. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/create-dataset-image.test.js:25:28) at Context.<anonymous> (test/create-dataset-image.test.js:50:5) at processImmediate (internal/timers.js:461:21)</pre></details>
priority
ai platform create dataset image after all hook should delete the created dataset for should create a new image dataset in the parent resource failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output command failed node delete dataset js undefined undefined undefined permission denied permission denied consumer project undefined has been suspended error command failed node delete dataset js undefined undefined undefined permission denied permission denied consumer project undefined has been suspended at checkexecsyncerror child process js at object execsync child process js at execsync test create dataset image test js at context test create dataset image test js at processimmediate internal timers js
1
16,404
2,614,983,211
IssuesEvent
2015-03-01 02:08:12
metapolator/metapolator
https://api.github.com/repos/metapolator/metapolator
opened
Implement sizing rules
Priority Low UI
https://github.com/metapolator/metapolator/wiki/elements-of-design#sizing-rules says, > **note**: all talk about pixels on these interaction design pages is about **effective pixels**, as reported in the browser (i.e. divided down from physical pixels for retina displays). > > 1. the default column and divider proportions are giving for each view above; > * users can adjust the column widths and divider proportions; this will not cause the views to scroll, and is persisted for the project; resizing the masters or instances column will affect both views they appear on; > * when the browser viewport is resized, column and divider proportions are maintained; > * **all three** rules above are subject to the following limits: > * minimum width of a view (what is shown in the viewport) is 960px (i.e. if the viewport is narrower than that, the view scrolls horizontally); > * the minimum width of the parameter, masters and instances columns is 180px; > * the maximum width of the parameter, masters and instances columns is 270px; > * the fonts column width is not user adjustable, it is 1/16th of the viewport width, capped by a maximum of 80px; > * no further limits; user common sense prevails. > > _yes, horizontal dividers can be moved completely up/down, to the point where the panel above/below it is completely invisible._ From this I see these tasks for the demo: - [ ] make the column widths adjustable; this will not cause the views to scroll - [ ] make the dividers adjustable; this will not cause the views to scroll - [ ] when the browser viewport is resized, column and divider proportions are maintained - [ ] minimum width of a view (what is shown in the viewport) is 960px (i.e. if the viewport is narrower than that, the view scrolls horizontally) - [ ] horizontal dividers can be moved completely up/down, to the point where the panel above/below it is completely invisible - [ ] the minimum width of the parameter, masters and instances columns is 180px; - [ ] the maximum width of the parameter, masters and instances columns is 270px; And from playing around, I add this as a 'nice to have' - [ ] when the browser viewport is resized, one of the 3 buttons to move from page to page must be pressed to re-align the panels to the viewport; it should be easy to fire the event triggered by these buttons on a resize event with the button's state as input, so that after the viewport resize operation is completed by the user, the UI slides back into alignment smoothly :) Assigning to Peter to verify - please edit the above and then assign to Jeroen when happy :)
1.0
Implement sizing rules - https://github.com/metapolator/metapolator/wiki/elements-of-design#sizing-rules says, > **note**: all talk about pixels on these interaction design pages is about **effective pixels**, as reported in the browser (i.e. divided down from physical pixels for retina displays). > > 1. the default column and divider proportions are giving for each view above; > * users can adjust the column widths and divider proportions; this will not cause the views to scroll, and is persisted for the project; resizing the masters or instances column will affect both views they appear on; > * when the browser viewport is resized, column and divider proportions are maintained; > * **all three** rules above are subject to the following limits: > * minimum width of a view (what is shown in the viewport) is 960px (i.e. if the viewport is narrower than that, the view scrolls horizontally); > * the minimum width of the parameter, masters and instances columns is 180px; > * the maximum width of the parameter, masters and instances columns is 270px; > * the fonts column width is not user adjustable, it is 1/16th of the viewport width, capped by a maximum of 80px; > * no further limits; user common sense prevails. > > _yes, horizontal dividers can be moved completely up/down, to the point where the panel above/below it is completely invisible._ From this I see these tasks for the demo: - [ ] make the column widths adjustable; this will not cause the views to scroll - [ ] make the dividers adjustable; this will not cause the views to scroll - [ ] when the browser viewport is resized, column and divider proportions are maintained - [ ] minimum width of a view (what is shown in the viewport) is 960px (i.e. if the viewport is narrower than that, the view scrolls horizontally) - [ ] horizontal dividers can be moved completely up/down, to the point where the panel above/below it is completely invisible - [ ] the minimum width of the parameter, masters and instances columns is 180px; - [ ] the maximum width of the parameter, masters and instances columns is 270px; And from playing around, I add this as a 'nice to have' - [ ] when the browser viewport is resized, one of the 3 buttons to move from page to page must be pressed to re-align the panels to the viewport; it should be easy to fire the event triggered by these buttons on a resize event with the button's state as input, so that after the viewport resize operation is completed by the user, the UI slides back into alignment smoothly :) Assigning to Peter to verify - please edit the above and then assign to Jeroen when happy :)
priority
implement sizing rules says note all talk about pixels on these interaction design pages is about effective pixels as reported in the browser i e divided down from physical pixels for retina displays the default column and divider proportions are giving for each view above users can adjust the column widths and divider proportions this will not cause the views to scroll and is persisted for the project resizing the masters or instances column will affect both views they appear on when the browser viewport is resized column and divider proportions are maintained all three rules above are subject to the following limits minimum width of a view what is shown in the viewport is i e if the viewport is narrower than that the view scrolls horizontally the minimum width of the parameter masters and instances columns is the maximum width of the parameter masters and instances columns is the fonts column width is not user adjustable it is of the viewport width capped by a maximum of no further limits user common sense prevails yes horizontal dividers can be moved completely up down to the point where the panel above below it is completely invisible from this i see these tasks for the demo make the column widths adjustable this will not cause the views to scroll make the dividers adjustable this will not cause the views to scroll when the browser viewport is resized column and divider proportions are maintained minimum width of a view what is shown in the viewport is i e if the viewport is narrower than that the view scrolls horizontally horizontal dividers can be moved completely up down to the point where the panel above below it is completely invisible the minimum width of the parameter masters and instances columns is the maximum width of the parameter masters and instances columns is and from playing around i add this as a nice to have when the browser viewport is resized one of the buttons to move from page to page must be pressed to re align the panels to the viewport it should be easy to fire the event triggered by these buttons on a resize event with the button s state as input so that after the viewport resize operation is completed by the user the ui slides back into alignment smoothly assigning to peter to verify please edit the above and then assign to jeroen when happy
1
113,365
4,552,972,677
IssuesEvent
2016-09-13 01:49:34
deis/controller
https://api.github.com/repos/deis/controller
closed
'deis pull' not using Procfile, causes 404 and 503 errors when using deis ps:scale
bug in progress priority 1
@helgi @kmala following up from our earlier conversation: `Procfile` doesn't seemed to be considered when deploying an app image via `deis pull`. This is on v2.4.0. STR: 1. `deis create odoyle-rules` 1. `deis config:set BUNCH_OF_ENV_VARS ... -a odoyle-rules` 1. `deis registry:set username=... password=... -a odoyle-rules` 1. `deis pull quay.io/... -a odoyle-rules` 1. odoyle-rules.example.com works at this point and is running a single `cmd` process. 1. Attempting to scale processes to match `Procfile` in the project yields a 404 error: `deis ps:scale cmd=0 web=1 worker=1` ``` $ deis ps:scale cmd=0 web=1 worker=1 -a odoyle-rules Scaling processes... but first, coffee! Error: Not Found ## controller [2016-09-12T22:23:53+00:00] - deis/deis-controller - 172.20.75.70 - - - 404 - "POST /v2/apps/odoyle-rules/scale/ HTTP/1.1" - 522 - "-" - "Deis Client vv2.4.0" - "~^deis\x5C.(?<domain>.+)$" - 100.66.35.56:80 - deis.example.com - 0.272 - 0.366 ``` 1. Using a `git push odoyle-rules master` deploy fixes the 404 and allows `deis ps:scale` work: ``` $ deis ps:scale web=1 worker=1 -a odoyle-rules Scaling processes... but first, coffee! done in 31s === odoyle-rules Processes --- web: odoyle-rules-web-3986992675-t34p4 up (v6) --- worker: odoyle-rules-worker-4178919286-eex5b up (v6) ``` However the router is set to permanently return a 503: ``` server { listen 8080; server_name ~^odoyle-rules\.(?<domain>.+)$; server_name_in_redirect off; port_in_redirect off; set $app_name "odoyle-rules"; listen 6443 ssl http2 ; ssl_protocols TLSv1 TLSv1.1 TLSv1.2; ssl_ciphers ECDHE-ECDSA-CHACHA20-POLY1305:ECDHE-RSA-CHACHA20-POLY1305:ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:ECDHE-ECDSA-AES128-SHA256:ECDHE-RSA-AES128-SHA256:ECDHE-ECDSA-AES128-SHA:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES128-SHA:ECDHE-ECDSA-AES256-SHA384:ECDHE-ECDSA-AES256-SHA:ECDHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA256:DHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA256:DHE-RSA-AES256-SHA:ECDHE-ECDSA-DES-CBC3-SHA:ECDHE-RSA-DES-CBC3-SHA:EDH-RSA-DES-CBC3-SHA:AES128-GCM-SHA256:AES256-GCM-SHA384:AES128-SHA256:AES256-SHA256:AES128-SHA:AES256-SHA:DES-CBC3-SHA:!DSS; ssl_prefer_server_ciphers on; ssl_certificate /opt/router/ssl/odoyle-rules.crt; ssl_certificate_key /opt/router/ssl/odoyle-rules.key; ssl_session_tickets on; ssl_buffer_size 4k; ssl_dhparam /opt/router/ssl/dhparam.pem; vhost_traffic_status_filter_by_set_key odoyle-rules application::*; location / { return 503; } } ```
1.0
'deis pull' not using Procfile, causes 404 and 503 errors when using deis ps:scale - @helgi @kmala following up from our earlier conversation: `Procfile` doesn't seemed to be considered when deploying an app image via `deis pull`. This is on v2.4.0. STR: 1. `deis create odoyle-rules` 1. `deis config:set BUNCH_OF_ENV_VARS ... -a odoyle-rules` 1. `deis registry:set username=... password=... -a odoyle-rules` 1. `deis pull quay.io/... -a odoyle-rules` 1. odoyle-rules.example.com works at this point and is running a single `cmd` process. 1. Attempting to scale processes to match `Procfile` in the project yields a 404 error: `deis ps:scale cmd=0 web=1 worker=1` ``` $ deis ps:scale cmd=0 web=1 worker=1 -a odoyle-rules Scaling processes... but first, coffee! Error: Not Found ## controller [2016-09-12T22:23:53+00:00] - deis/deis-controller - 172.20.75.70 - - - 404 - "POST /v2/apps/odoyle-rules/scale/ HTTP/1.1" - 522 - "-" - "Deis Client vv2.4.0" - "~^deis\x5C.(?<domain>.+)$" - 100.66.35.56:80 - deis.example.com - 0.272 - 0.366 ``` 1. Using a `git push odoyle-rules master` deploy fixes the 404 and allows `deis ps:scale` work: ``` $ deis ps:scale web=1 worker=1 -a odoyle-rules Scaling processes... but first, coffee! done in 31s === odoyle-rules Processes --- web: odoyle-rules-web-3986992675-t34p4 up (v6) --- worker: odoyle-rules-worker-4178919286-eex5b up (v6) ``` However the router is set to permanently return a 503: ``` server { listen 8080; server_name ~^odoyle-rules\.(?<domain>.+)$; server_name_in_redirect off; port_in_redirect off; set $app_name "odoyle-rules"; listen 6443 ssl http2 ; ssl_protocols TLSv1 TLSv1.1 TLSv1.2; ssl_ciphers ECDHE-ECDSA-CHACHA20-POLY1305:ECDHE-RSA-CHACHA20-POLY1305:ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384:DHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384:ECDHE-ECDSA-AES128-SHA256:ECDHE-RSA-AES128-SHA256:ECDHE-ECDSA-AES128-SHA:ECDHE-RSA-AES256-SHA384:ECDHE-RSA-AES128-SHA:ECDHE-ECDSA-AES256-SHA384:ECDHE-ECDSA-AES256-SHA:ECDHE-RSA-AES256-SHA:DHE-RSA-AES128-SHA256:DHE-RSA-AES128-SHA:DHE-RSA-AES256-SHA256:DHE-RSA-AES256-SHA:ECDHE-ECDSA-DES-CBC3-SHA:ECDHE-RSA-DES-CBC3-SHA:EDH-RSA-DES-CBC3-SHA:AES128-GCM-SHA256:AES256-GCM-SHA384:AES128-SHA256:AES256-SHA256:AES128-SHA:AES256-SHA:DES-CBC3-SHA:!DSS; ssl_prefer_server_ciphers on; ssl_certificate /opt/router/ssl/odoyle-rules.crt; ssl_certificate_key /opt/router/ssl/odoyle-rules.key; ssl_session_tickets on; ssl_buffer_size 4k; ssl_dhparam /opt/router/ssl/dhparam.pem; vhost_traffic_status_filter_by_set_key odoyle-rules application::*; location / { return 503; } } ```
priority
deis pull not using procfile causes and errors when using deis ps scale helgi kmala following up from our earlier conversation procfile doesn t seemed to be considered when deploying an app image via deis pull this is on str deis create odoyle rules deis config set bunch of env vars a odoyle rules deis registry set username password a odoyle rules deis pull quay io a odoyle rules odoyle rules example com works at this point and is running a single cmd process attempting to scale processes to match procfile in the project yields a error deis ps scale cmd web worker deis ps scale cmd web worker a odoyle rules scaling processes but first coffee error not found controller deis deis controller post apps odoyle rules scale http deis client deis deis example com using a git push odoyle rules master deploy fixes the and allows deis ps scale work deis ps scale web worker a odoyle rules scaling processes but first coffee done in odoyle rules processes web odoyle rules web up worker odoyle rules worker up however the router is set to permanently return a server listen server name odoyle rules server name in redirect off port in redirect off set app name odoyle rules listen ssl ssl protocols ssl ciphers ecdhe ecdsa ecdhe rsa ecdhe ecdsa gcm ecdhe rsa gcm ecdhe ecdsa gcm ecdhe rsa gcm dhe rsa gcm dhe rsa gcm ecdhe ecdsa ecdhe rsa ecdhe ecdsa sha ecdhe rsa ecdhe rsa sha ecdhe ecdsa ecdhe ecdsa sha ecdhe rsa sha dhe rsa dhe rsa sha dhe rsa dhe rsa sha ecdhe ecdsa des sha ecdhe rsa des sha edh rsa des sha gcm gcm sha sha des sha dss ssl prefer server ciphers on ssl certificate opt router ssl odoyle rules crt ssl certificate key opt router ssl odoyle rules key ssl session tickets on ssl buffer size ssl dhparam opt router ssl dhparam pem vhost traffic status filter by set key odoyle rules application location return
1
383,928
26,569,478,148
IssuesEvent
2023-01-21 01:08:09
ClydiesF/Hurd
https://api.github.com/repos/ClydiesF/Hurd
closed
Create A Awesome Project README for Hurd
documentation good first issue
the Readme for inspire someone to collaborate and explain exactly what Hurd is here for and what it aims to accomplish.
1.0
Create A Awesome Project README for Hurd - the Readme for inspire someone to collaborate and explain exactly what Hurd is here for and what it aims to accomplish.
non_priority
create a awesome project readme for hurd the readme for inspire someone to collaborate and explain exactly what hurd is here for and what it aims to accomplish
0
14,189
4,836,333,310
IssuesEvent
2016-11-08 19:18:54
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Why /administrator/includes/helper.php is stiil there?
No Code Attached Yet
### Steps to reproduce the issue Head to /administrator/includes/ helper.php is there. Shouldnt be there as its deprecated. ### Expected result Removal of said file
1.0
Why /administrator/includes/helper.php is stiil there? - ### Steps to reproduce the issue Head to /administrator/includes/ helper.php is there. Shouldnt be there as its deprecated. ### Expected result Removal of said file
non_priority
why administrator includes helper php is stiil there steps to reproduce the issue head to administrator includes helper php is there shouldnt be there as its deprecated expected result removal of said file
0
7,417
2,852,108,916
IssuesEvent
2015-06-01 11:39:03
bikeanjo/bikeanjo
https://api.github.com/repos/bikeanjo/bikeanjo
opened
Parralax - Imagens de Background
design efeitos / animação
@ariadnemelo , um tempo atrás eu tinha comentando que gostaria de incluir alguns efeitos de transição e animação no site, por isso criei uma label nova chamada "efeitos / animação" para ir indicando onde seria bacana ter algum efeito e como. Lembrando que esses efeitos são conversáveis, principalmente se impactarem muito o andamento do projeto, a gente pode ir conversando para ir vendo qual é viável ou não, blz? O primeiro seria nas imagens de background da home e das páginas estáticas, colocar um efeito parallax, como esses exemplos: https://github.com/IanLunn/jQuery-Parallax https://github.com/pixelcog/parallax.js/
1.0
Parralax - Imagens de Background - @ariadnemelo , um tempo atrás eu tinha comentando que gostaria de incluir alguns efeitos de transição e animação no site, por isso criei uma label nova chamada "efeitos / animação" para ir indicando onde seria bacana ter algum efeito e como. Lembrando que esses efeitos são conversáveis, principalmente se impactarem muito o andamento do projeto, a gente pode ir conversando para ir vendo qual é viável ou não, blz? O primeiro seria nas imagens de background da home e das páginas estáticas, colocar um efeito parallax, como esses exemplos: https://github.com/IanLunn/jQuery-Parallax https://github.com/pixelcog/parallax.js/
non_priority
parralax imagens de background ariadnemelo um tempo atrás eu tinha comentando que gostaria de incluir alguns efeitos de transição e animação no site por isso criei uma label nova chamada efeitos animação para ir indicando onde seria bacana ter algum efeito e como lembrando que esses efeitos são conversáveis principalmente se impactarem muito o andamento do projeto a gente pode ir conversando para ir vendo qual é viável ou não blz o primeiro seria nas imagens de background da home e das páginas estáticas colocar um efeito parallax como esses exemplos
0
712,899
24,510,351,932
IssuesEvent
2022-10-10 20:43:17
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Prepare rancher-webhook chart for 2.7.0
kind/enhancement priority/0 [zube]: QA Working QA/XS team/area1 area/charts
rancher-webhook chart needs to be updated for 2.7 by 1) File a PR removing the `doNotRelease` flag form the package.yaml; in this PR, you will also want to make any changes that need to be made to update the chart to the version you intend to release in 2.7. For example, for a charts like rancher-aks-operator / rancher-gke-operator / rancher-eks-operator, you will need to create a new branch in upstream for Rancher 2.7, cut the first release for that chart, and introduce those changes in to the package.yaml as well before removing the doNotRelease flag 2) To make your chart available for 2.7, almost every chart will also have to modify the Rancher Version annotations that scope it to 2.6, i.e. https://github.com/rancher/charts/blob/2e058e93ee01f2980e4750148842dd7f04262b38/charts/fleet-agent/100.0.5%2Bup0.3.11/Chart.yaml#L8 3) Once the PR is merged, ensure that the brand new chart is smoke tested by QA/dev
1.0
Prepare rancher-webhook chart for 2.7.0 - rancher-webhook chart needs to be updated for 2.7 by 1) File a PR removing the `doNotRelease` flag form the package.yaml; in this PR, you will also want to make any changes that need to be made to update the chart to the version you intend to release in 2.7. For example, for a charts like rancher-aks-operator / rancher-gke-operator / rancher-eks-operator, you will need to create a new branch in upstream for Rancher 2.7, cut the first release for that chart, and introduce those changes in to the package.yaml as well before removing the doNotRelease flag 2) To make your chart available for 2.7, almost every chart will also have to modify the Rancher Version annotations that scope it to 2.6, i.e. https://github.com/rancher/charts/blob/2e058e93ee01f2980e4750148842dd7f04262b38/charts/fleet-agent/100.0.5%2Bup0.3.11/Chart.yaml#L8 3) Once the PR is merged, ensure that the brand new chart is smoke tested by QA/dev
priority
prepare rancher webhook chart for rancher webhook chart needs to be updated for by file a pr removing the donotrelease flag form the package yaml in this pr you will also want to make any changes that need to be made to update the chart to the version you intend to release in for example for a charts like rancher aks operator rancher gke operator rancher eks operator you will need to create a new branch in upstream for rancher cut the first release for that chart and introduce those changes in to the package yaml as well before removing the donotrelease flag to make your chart available for almost every chart will also have to modify the rancher version annotations that scope it to i e once the pr is merged ensure that the brand new chart is smoke tested by qa dev
1
470,094
13,530,856,014
IssuesEvent
2020-09-15 20:36:34
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID :211049] Macro compares unsigned to 0 in arch/x86/core/x86_mmu.c
Coverity bug priority: low
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/8e2c4a475dc375da6691175dd1da87525053ed76/arch/x86/core/x86_mmu.c#L673 Category: Integer handling issues Function: `is_within_system_ram` Component: Architectures CID: [211049](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=211049) Details: ``` 667 static inline bool is_within_system_ram(uintptr_t addr) 668 { 669 #ifdef CONFIG_X86_64 670 /* FIXME: locore not included in CONFIG_SRAM_BASE_ADDRESS */ 671 return addr < (PHYS_RAM_ADDR + PHYS_RAM_SIZE); 672 #else >>> CID 211049: Integer handling issues (NO_EFFECT) >>> This greater-than-or-equal-to-zero comparison of an unsigned value is always true. "addr >= 0UL". 673 return (addr >= PHYS_RAM_ADDR) && 674 (addr < (PHYS_RAM_ADDR + PHYS_RAM_SIZE)); 675 #endif 676 } 677 678 /* Ignored bit posiition at all levels */ ``` Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID :211049] Macro compares unsigned to 0 in arch/x86/core/x86_mmu.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/8e2c4a475dc375da6691175dd1da87525053ed76/arch/x86/core/x86_mmu.c#L673 Category: Integer handling issues Function: `is_within_system_ram` Component: Architectures CID: [211049](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=211049) Details: ``` 667 static inline bool is_within_system_ram(uintptr_t addr) 668 { 669 #ifdef CONFIG_X86_64 670 /* FIXME: locore not included in CONFIG_SRAM_BASE_ADDRESS */ 671 return addr < (PHYS_RAM_ADDR + PHYS_RAM_SIZE); 672 #else >>> CID 211049: Integer handling issues (NO_EFFECT) >>> This greater-than-or-equal-to-zero comparison of an unsigned value is always true. "addr >= 0UL". 673 return (addr >= PHYS_RAM_ADDR) && 674 (addr < (PHYS_RAM_ADDR + PHYS_RAM_SIZE)); 675 #endif 676 } 677 678 /* Ignored bit posiition at all levels */ ``` Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
priority
macro compares unsigned to in arch core mmu c static code scan issues found in file category integer handling issues function is within system ram component architectures cid details static inline bool is within system ram uintptr t addr ifdef config fixme locore not included in config sram base address return addr phys ram addr phys ram size else cid integer handling issues no effect this greater than or equal to zero comparison of an unsigned value is always true addr return addr phys ram addr addr phys ram addr phys ram size endif ignored bit posiition at all levels please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
1
111,529
14,105,900,438
IssuesEvent
2020-11-06 14:10:59
reactor/reactor-core
https://api.github.com/repos/reactor/reactor-core
closed
Reduce the number of resolved currentContext() where possible
area/performance for/team-attention status/declined status/need-design type/enhancement
Somehow related to #1011 but in the meantime we need to scale down the number of actualContext resolution we do like in FluxConcatMap etc. In the scrabble benchmarks, it does seem to incur a 10-15% perf penalty.
1.0
Reduce the number of resolved currentContext() where possible - Somehow related to #1011 but in the meantime we need to scale down the number of actualContext resolution we do like in FluxConcatMap etc. In the scrabble benchmarks, it does seem to incur a 10-15% perf penalty.
non_priority
reduce the number of resolved currentcontext where possible somehow related to but in the meantime we need to scale down the number of actualcontext resolution we do like in fluxconcatmap etc in the scrabble benchmarks it does seem to incur a perf penalty
0
121,826
16,041,217,500
IssuesEvent
2021-04-22 08:08:26
mrdoob/three.js
https://api.github.com/repos/mrdoob/three.js
closed
Make WebXR tree-shakable
Design
The current implementation/design of `WebXRManager` (12kb uncompressed) is not tree-shakeable. Current rollup-bundle-visualizer: ![](https://aws1.discourse-cdn.com/standard17/uploads/threejs/original/2X/e/ea9cd6e6caa53773c23fec6ed09fbc961ad45622.png)
1.0
Make WebXR tree-shakable - The current implementation/design of `WebXRManager` (12kb uncompressed) is not tree-shakeable. Current rollup-bundle-visualizer: ![](https://aws1.discourse-cdn.com/standard17/uploads/threejs/original/2X/e/ea9cd6e6caa53773c23fec6ed09fbc961ad45622.png)
non_priority
make webxr tree shakable the current implementation design of webxrmanager uncompressed is not tree shakeable current rollup bundle visualizer
0
311,928
9,540,588,217
IssuesEvent
2019-04-30 19:56:29
zeit/ncc
https://api.github.com/repos/zeit/ncc
reopened
osapi fails to build (depends on noda)
package issue priority
Hello, I cannot deploy my backend lambda function because of strange error and I want to ask for your help if it's possible. General issue is I'm using [`osapi`](https://www.npmjs.com/package/osapi) for my s3 connections and stuff and this package works well while I'm working locally but since it's deployed backend sends 502 on any lambda where piece of code with requiring osapi is present. I created a public deployment: [https://zeit.co/kkz-zae/express/g081kykjq ](https://zeit.co/kkz-zae/express/g081kykjq ) and also GitHub repository to reproduce this: [https://github.com/reetou/now-reproduce-express](https://github.com/reetou/now-reproduce-express) As far as I understand it cannot find some dependencies of osapi for some reason. Hope you will help me or feed my interest why it happened at least.
1.0
osapi fails to build (depends on noda) - Hello, I cannot deploy my backend lambda function because of strange error and I want to ask for your help if it's possible. General issue is I'm using [`osapi`](https://www.npmjs.com/package/osapi) for my s3 connections and stuff and this package works well while I'm working locally but since it's deployed backend sends 502 on any lambda where piece of code with requiring osapi is present. I created a public deployment: [https://zeit.co/kkz-zae/express/g081kykjq ](https://zeit.co/kkz-zae/express/g081kykjq ) and also GitHub repository to reproduce this: [https://github.com/reetou/now-reproduce-express](https://github.com/reetou/now-reproduce-express) As far as I understand it cannot find some dependencies of osapi for some reason. Hope you will help me or feed my interest why it happened at least.
priority
osapi fails to build depends on noda hello i cannot deploy my backend lambda function because of strange error and i want to ask for your help if it s possible general issue is i m using for my connections and stuff and this package works well while i m working locally but since it s deployed backend sends on any lambda where piece of code with requiring osapi is present i created a public deployment and also github repository to reproduce this as far as i understand it cannot find some dependencies of osapi for some reason hope you will help me or feed my interest why it happened at least
1
249,273
21,156,438,641
IssuesEvent
2022-04-07 04:09:41
enpaul/tox-poetry-installer
https://api.github.com/repos/enpaul/tox-poetry-installer
closed
Add integration tests
roadmap-stable tests
A series of integration tests should be added (in addition to the unit tests already present). The goal of the integration tests is to validate the integration with Tox, since that component of the program can't be adequately tested via unit tests because Tox is not directly imported anywhere. Emphasis of the integration tests should be on the various config options having the intended effects and the proper dependencies being "installed" to the venv at runtime.
1.0
Add integration tests - A series of integration tests should be added (in addition to the unit tests already present). The goal of the integration tests is to validate the integration with Tox, since that component of the program can't be adequately tested via unit tests because Tox is not directly imported anywhere. Emphasis of the integration tests should be on the various config options having the intended effects and the proper dependencies being "installed" to the venv at runtime.
non_priority
add integration tests a series of integration tests should be added in addition to the unit tests already present the goal of the integration tests is to validate the integration with tox since that component of the program can t be adequately tested via unit tests because tox is not directly imported anywhere emphasis of the integration tests should be on the various config options having the intended effects and the proper dependencies being installed to the venv at runtime
0
1,175
2,615,122,642
IssuesEvent
2015-03-01 05:49:56
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
Inner classes for unnamed objects in generated Java library
auto-migrated Component-Google-APIs Milestone-CodeGenVersion1.4.0 Priority-Medium Type-Enhancement
``` Currently we generate top-level classes like this: http://javadoc.google-api-java-client.googlecode.com/hg/apis/discovery/com/googl e/api/services/discovery/model/RestDescriptionAuthOauth2Scopes.html I am not happy with two aspects of this: 1. Class name is too long. It should be simpler. 2. It clutters up the top-level class names list: http://javadoc.google-api-java-client.googlecode.com/hg/apis/discovery/com/googl e/api/services/discovery/model/package-frame.html Instead, if we generated these as inner classes, it would solve both of these issues. So it should instead generate something like this: public class RestDescription { public static class Auth { public static class Oauth2 { public static class Scopes { ... Note that this way I can declare the import like this: import com.google.api.services.discovery.model.RestDescription.Auth.Oauth2.Scopes; and then refer to it simply as Scopes in my code. ``` Original issue reported on code.google.com by `rmis...@google.com` on 27 Sep 2011 at 3:46
1.0
Inner classes for unnamed objects in generated Java library - ``` Currently we generate top-level classes like this: http://javadoc.google-api-java-client.googlecode.com/hg/apis/discovery/com/googl e/api/services/discovery/model/RestDescriptionAuthOauth2Scopes.html I am not happy with two aspects of this: 1. Class name is too long. It should be simpler. 2. It clutters up the top-level class names list: http://javadoc.google-api-java-client.googlecode.com/hg/apis/discovery/com/googl e/api/services/discovery/model/package-frame.html Instead, if we generated these as inner classes, it would solve both of these issues. So it should instead generate something like this: public class RestDescription { public static class Auth { public static class Oauth2 { public static class Scopes { ... Note that this way I can declare the import like this: import com.google.api.services.discovery.model.RestDescription.Auth.Oauth2.Scopes; and then refer to it simply as Scopes in my code. ``` Original issue reported on code.google.com by `rmis...@google.com` on 27 Sep 2011 at 3:46
non_priority
inner classes for unnamed objects in generated java library currently we generate top level classes like this e api services discovery model html i am not happy with two aspects of this class name is too long it should be simpler it clutters up the top level class names list e api services discovery model package frame html instead if we generated these as inner classes it would solve both of these issues so it should instead generate something like this public class restdescription public static class auth public static class public static class scopes note that this way i can declare the import like this import com google api services discovery model restdescription auth scopes and then refer to it simply as scopes in my code original issue reported on code google com by rmis google com on sep at
0
447,975
31,759,476,172
IssuesEvent
2023-09-12 03:09:01
allzeroyou/Algorithm
https://api.github.com/repos/allzeroyou/Algorithm
opened
BOJ 10845 큐 풀이
documentation
## 문제 분석 ### 첫 번째 단계(문제 요약 및 조건 파악) 정수를 저장하는 큐 구현하고, 입력으로 주어지는 명령 처리 명령은 총 6가지 - push x: 정수 x를 큐에 넣음 - pop: 큐에서 가장 앞에 있는 정수 꺼내기, 그 수를 출력. 만약 큐가 비었다면 -1 출력 - size: 큐에 들어있는 정수 개수 출력 - empty: 큐가 비어있으면 1, 아니면 0 출력 - front: 큐에서 가장 앞에 있는 정수 출력. 만약 큐가 비었다면 -1 출력 - back: 큐에서 가장 뒤에 있는 정수 출력. 만약 큐가 비었다면 -1 출력 - 입력 첫째 줄에 주어지는 명령의 수 N (1 ≤ N ≤ 10,000)이 주어짐. 둘째 줄부터 N개의 줄에는 `명령`이 하나씩 주어진다 주어지는 정수는 1보다 크거나 같고, 100,000보다 작거나 같다. 문제에 나와있지 않은 명령이 주어지는 경우는 없다. - 출력 출력해야하는 명령이 주어질 때마다, 한 줄에 하나씩 출력한다. ### 두 번째 단계 (문제 핵심 파악) 6가지 기능을 갖춘 `큐`를 구현한다 파이썬에서 제공하는 큐 자료구조인 `deque`를 이용하자 - **큐 구현시 list를 이용하지 않는 이유** 스택에서 list.append와 list.pop()을 이용했던 것처럼 list.append와 list.pop(0)을 이용하면 리스트를 큐처럼 사용할 수 있다. 하지만 pop()의 time complexity는 O(1)인 반면 pop(0)의 time complexity는 O(N)이기 때문에 시간이 오래 걸린다. 따라서 시간 복잡도를 고려해 리스트는 큐로 사용하지 않는다. ## 코드 작성 ```python import collections import sys input = sys.stdin.readline q = collections.deque() # 큐 생성 n = int(input()) for _ in range(n): command = input().split() if command[0] == "push": # push 명령 q.append(command[1]) elif command[0] == "front": if not q: # 큐가 비었다면 print(-1) else: print(q[0]) elif command[0] == "back": if not q: # 큐가 비었다면 print(-1) else: print(q[-1]) elif command[0] == "size": print(len(q)) elif command[0] == "empty": if not q: # 큐가 비었다면 print(1) else: print(0) elif command[0] == "pop": if not q: # 큐가 비었다면 print(-1) else: print(q.popleft()) ``` ### 느낀점 <p><strong>deque vs heapq</strong></p> <p>deque는 스택+큐 자료구조. 가장자리의 원소를 넣거나 뺄 수 있다. pop()과 popleft() 모두 시간복잡도가 O(1)로 매우 좋다.</p> 메서드 | 설명 -- | -- deque(iterable, [, maxlen]) | 초기화 함수이다. iterable(리스트 등)을 인자로 건내면 이를 deque화 시켜준다. append(x) | x를 덱의 오른쪽에 삽입한다. popleft() | 덱의 가장 왼쪽에 있는 원소를 덱에서 제거하고, 그 값을 리턴한다. clear() | 모든 원소를 지운다. <p>heapq는 우선순위 큐. 최소 힙을 지원하며, 최단 경로를 탐색하는 다익스트라 알고리즘 등에 사용됨.</p> <ul> <li>근데 <strong>힙</strong>은 뭔데? <ul> <li> <p><strong>최솟값, 최댓값</strong>을 빠르게 찾기 위해 고안된 <strong>완전 이진 트리</strong>.</p> </li> <li> <p>최솟값이나 최댓값을 찾기 위해 배열을 사용하면 Ο(n)만큼 시간이 걸린다.</p> <p>하지만 힙을 사용하면 O(logn)만큼 소요되므로, 배열을 사용할 때보다 빠르게 최솟값과 최댓값을 구할 수 있다.</p> <p>우선순위 큐와 같이 최댓값 또는 최솟값을 빠르게 찾아야하는 알고리즘 등에 활용된다.</p> </li> </ul> </li> </ul> 참고 블로그 - deque (https://ooeunz.tistory.com/31) - heap (https://velog.io/@gnwjd309/data-structure-heap)
1.0
BOJ 10845 큐 풀이 - ## 문제 분석 ### 첫 번째 단계(문제 요약 및 조건 파악) 정수를 저장하는 큐 구현하고, 입력으로 주어지는 명령 처리 명령은 총 6가지 - push x: 정수 x를 큐에 넣음 - pop: 큐에서 가장 앞에 있는 정수 꺼내기, 그 수를 출력. 만약 큐가 비었다면 -1 출력 - size: 큐에 들어있는 정수 개수 출력 - empty: 큐가 비어있으면 1, 아니면 0 출력 - front: 큐에서 가장 앞에 있는 정수 출력. 만약 큐가 비었다면 -1 출력 - back: 큐에서 가장 뒤에 있는 정수 출력. 만약 큐가 비었다면 -1 출력 - 입력 첫째 줄에 주어지는 명령의 수 N (1 ≤ N ≤ 10,000)이 주어짐. 둘째 줄부터 N개의 줄에는 `명령`이 하나씩 주어진다 주어지는 정수는 1보다 크거나 같고, 100,000보다 작거나 같다. 문제에 나와있지 않은 명령이 주어지는 경우는 없다. - 출력 출력해야하는 명령이 주어질 때마다, 한 줄에 하나씩 출력한다. ### 두 번째 단계 (문제 핵심 파악) 6가지 기능을 갖춘 `큐`를 구현한다 파이썬에서 제공하는 큐 자료구조인 `deque`를 이용하자 - **큐 구현시 list를 이용하지 않는 이유** 스택에서 list.append와 list.pop()을 이용했던 것처럼 list.append와 list.pop(0)을 이용하면 리스트를 큐처럼 사용할 수 있다. 하지만 pop()의 time complexity는 O(1)인 반면 pop(0)의 time complexity는 O(N)이기 때문에 시간이 오래 걸린다. 따라서 시간 복잡도를 고려해 리스트는 큐로 사용하지 않는다. ## 코드 작성 ```python import collections import sys input = sys.stdin.readline q = collections.deque() # 큐 생성 n = int(input()) for _ in range(n): command = input().split() if command[0] == "push": # push 명령 q.append(command[1]) elif command[0] == "front": if not q: # 큐가 비었다면 print(-1) else: print(q[0]) elif command[0] == "back": if not q: # 큐가 비었다면 print(-1) else: print(q[-1]) elif command[0] == "size": print(len(q)) elif command[0] == "empty": if not q: # 큐가 비었다면 print(1) else: print(0) elif command[0] == "pop": if not q: # 큐가 비었다면 print(-1) else: print(q.popleft()) ``` ### 느낀점 <p><strong>deque vs heapq</strong></p> <p>deque는 스택+큐 자료구조. 가장자리의 원소를 넣거나 뺄 수 있다. pop()과 popleft() 모두 시간복잡도가 O(1)로 매우 좋다.</p> 메서드 | 설명 -- | -- deque(iterable, [, maxlen]) | 초기화 함수이다. iterable(리스트 등)을 인자로 건내면 이를 deque화 시켜준다. append(x) | x를 덱의 오른쪽에 삽입한다. popleft() | 덱의 가장 왼쪽에 있는 원소를 덱에서 제거하고, 그 값을 리턴한다. clear() | 모든 원소를 지운다. <p>heapq는 우선순위 큐. 최소 힙을 지원하며, 최단 경로를 탐색하는 다익스트라 알고리즘 등에 사용됨.</p> <ul> <li>근데 <strong>힙</strong>은 뭔데? <ul> <li> <p><strong>최솟값, 최댓값</strong>을 빠르게 찾기 위해 고안된 <strong>완전 이진 트리</strong>.</p> </li> <li> <p>최솟값이나 최댓값을 찾기 위해 배열을 사용하면 Ο(n)만큼 시간이 걸린다.</p> <p>하지만 힙을 사용하면 O(logn)만큼 소요되므로, 배열을 사용할 때보다 빠르게 최솟값과 최댓값을 구할 수 있다.</p> <p>우선순위 큐와 같이 최댓값 또는 최솟값을 빠르게 찾아야하는 알고리즘 등에 활용된다.</p> </li> </ul> </li> </ul> 참고 블로그 - deque (https://ooeunz.tistory.com/31) - heap (https://velog.io/@gnwjd309/data-structure-heap)
non_priority
boj 큐 풀이 문제 분석 첫 번째 단계 문제 요약 및 조건 파악 정수를 저장하는 큐 구현하고 입력으로 주어지는 명령 처리 명령은 총 push x 정수 x를 큐에 넣음 pop 큐에서 가장 앞에 있는 정수 꺼내기 그 수를 출력 만약 큐가 비었다면 출력 size 큐에 들어있는 정수 개수 출력 empty 큐가 비어있으면 아니면 출력 front 큐에서 가장 앞에 있는 정수 출력 만약 큐가 비었다면 출력 back 큐에서 가장 뒤에 있는 정수 출력 만약 큐가 비었다면 출력 입력 첫째 줄에 주어지는 명령의 수 n ≤ n ≤ 이 주어짐 둘째 줄부터 n개의 줄에는 명령 이 하나씩 주어진다 주어지는 정수는 크거나 같고 작거나 같다 문제에 나와있지 않은 명령이 주어지는 경우는 없다 출력 출력해야하는 명령이 주어질 때마다 한 줄에 하나씩 출력한다 두 번째 단계 문제 핵심 파악 기능을 갖춘 큐 를 구현한다 파이썬에서 제공하는 큐 자료구조인 deque 를 이용하자 큐 구현시 list를 이용하지 않는 이유 스택에서 list append와 list pop 을 이용했던 것처럼 list append와 list pop 을 이용하면 리스트를 큐처럼 사용할 수 있다 하지만 pop 의 time complexity는 o 인 반면 pop 의 time complexity는 o n 이기 때문에 시간이 오래 걸린다 따라서 시간 복잡도를 고려해 리스트는 큐로 사용하지 않는다 코드 작성 python import collections import sys input sys stdin readline q collections deque 큐 생성 n int input for in range n command input split if command push push 명령 q append command elif command front if not q 큐가 비었다면 print else print q elif command back if not q 큐가 비었다면 print else print q elif command size print len q elif command empty if not q 큐가 비었다면 print else print elif command pop if not q 큐가 비었다면 print else print q popleft 느낀점 deque vs heapq deque는 스택 큐 자료구조 가장자리의 원소를 넣거나 뺄 수 있다 pop 과 popleft 모두 시간복잡도가 o 로 매우 좋다 메서드 설명 deque iterable 초기화 함수이다 iterable 리스트 등 을 인자로 건내면 이를 deque화 시켜준다 append x x를 덱의 오른쪽에 삽입한다 popleft 덱의 가장 왼쪽에 있는 원소를 덱에서 제거하고 그 값을 리턴한다 clear 모든 원소를 지운다 heapq는 우선순위 큐 최소 힙을 지원하며 최단 경로를 탐색하는 다익스트라 알고리즘 등에 사용됨 근데 힙 은 뭔데 최솟값 최댓값 을 빠르게 찾기 위해 고안된 완전 이진 트리 최솟값이나 최댓값을 찾기 위해 배열을 사용하면 ο n 만큼 시간이 걸린다 하지만 힙을 사용하면 o logn 만큼 소요되므로 배열을 사용할 때보다 빠르게 최솟값과 최댓값을 구할 수 있다 우선순위 큐와 같이 최댓값 또는 최솟값을 빠르게 찾아야하는 알고리즘 등에 활용된다 참고 블로그 deque heap
0
29,610
2,716,631,617
IssuesEvent
2015-04-10 20:21:26
CruxFramework/crux
https://api.github.com/repos/CruxFramework/crux
closed
Error on client REST Proxies when @PathParam uses regular expression to restrict the parameter value
bug imported Milestone-M14-C3 Priority-Medium TargetVersion-5.1.3
_From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on August 27, 2014 16:14:13_ Create a REST method like: @Path("{id:\\d+}") @GET public String test(@PathParam("id") Integer id) { return "teste"; } Then try to invoke this method through a Crux RestProxy interface. The client proxies will not handle the regular expression correctly and will make a request to a wrong URI. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=498_
1.0
Error on client REST Proxies when @PathParam uses regular expression to restrict the parameter value - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on August 27, 2014 16:14:13_ Create a REST method like: @Path("{id:\\d+}") @GET public String test(@PathParam("id") Integer id) { return "teste"; } Then try to invoke this method through a Crux RestProxy interface. The client proxies will not handle the regular expression correctly and will make a request to a wrong URI. _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=498_
priority
error on client rest proxies when pathparam uses regular expression to restrict the parameter value from on august create a rest method like path id d get public string test pathparam id integer id return teste then try to invoke this method through a crux restproxy interface the client proxies will not handle the regular expression correctly and will make a request to a wrong uri original issue
1
784,322
27,566,394,389
IssuesEvent
2023-03-08 04:25:00
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[Query Expressions]Some queries fail resolving closure symbols
Type/Bug Priority/High Team/CompilerFE Area/Desugar Points/7 Lang/Expressions/Query Reason/EngineeringMistake Deferred
**Description:** ```ballerina var res = from var emp in (from var x in (from var xx in (from var xxx in (from var {firstName, lastName, dept} in employeeList where 1 == 1 select firstName + " " + lastName) select xxx) select xx) select x) select emp; ``` **Steps to reproduce:** **Affected Versions:** SL Beta 3 rc 3 **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
[Query Expressions]Some queries fail resolving closure symbols - **Description:** ```ballerina var res = from var emp in (from var x in (from var xx in (from var xxx in (from var {firstName, lastName, dept} in employeeList where 1 == 1 select firstName + " " + lastName) select xxx) select xx) select x) select emp; ``` **Steps to reproduce:** **Affected Versions:** SL Beta 3 rc 3 **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
priority
some queries fail resolving closure symbols description ballerina var res from var emp in from var x in from var xx in from var xxx in from var firstname lastname dept in employeelist where select firstname lastname select xxx select xx select x select emp steps to reproduce affected versions sl beta rc os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
1
789,812
27,806,530,561
IssuesEvent
2023-03-17 20:31:03
ScicraftLearn/Minelabs
https://api.github.com/repos/ScicraftLearn/Minelabs
closed
Schrödinger's Cat: Zombie noises
enhancement low priority shrodinger
When the box is unopened, it emits zombie-cat noises sporadically. For now the zombie noise can be used, perhaps with a higher pitch or combined with ocelot noises (do they make noise?).
1.0
Schrödinger's Cat: Zombie noises - When the box is unopened, it emits zombie-cat noises sporadically. For now the zombie noise can be used, perhaps with a higher pitch or combined with ocelot noises (do they make noise?).
priority
schrödinger s cat zombie noises when the box is unopened it emits zombie cat noises sporadically for now the zombie noise can be used perhaps with a higher pitch or combined with ocelot noises do they make noise
1
273,969
8,555,538,731
IssuesEvent
2018-11-08 10:19:55
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
estilysee.info - see bug description
browser-firefox priority-normal
<!-- @browser: Firefox 64.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://estilysee.info/ENZ?tag_id=715349&sub_id1=59&sub_id2=ADF69739-9A99-4A8D-9FBB-D6708D40FCE4&cookie_id=e8824c59-b7a0-4096-b093-b24653e75482&lp=blank&tb=redirect&allb=redirect&ob=redirect&href=https://qareaste.info/?tag_id=715349%26sub_id1=59%26sub_id2=ADF69739-9A99-4A8D-9FBB-D6708D40FCE4%26cookie_id=e8824c59-b7a0-4096-b093-b24653e75482%26lp=blank%26tb=redirect%26allb=redirect%26ob=redirect **Browser / Version**: Firefox 64.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: Site keeps popping up and I don't know what it is **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181101155334</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]', u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]', u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]', u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
estilysee.info - see bug description - <!-- @browser: Firefox 64.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://estilysee.info/ENZ?tag_id=715349&sub_id1=59&sub_id2=ADF69739-9A99-4A8D-9FBB-D6708D40FCE4&cookie_id=e8824c59-b7a0-4096-b093-b24653e75482&lp=blank&tb=redirect&allb=redirect&ob=redirect&href=https://qareaste.info/?tag_id=715349%26sub_id1=59%26sub_id2=ADF69739-9A99-4A8D-9FBB-D6708D40FCE4%26cookie_id=e8824c59-b7a0-4096-b093-b24653e75482%26lp=blank%26tb=redirect%26allb=redirect%26ob=redirect **Browser / Version**: Firefox 64.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: Site keeps popping up and I don't know what it is **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181101155334</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]', u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]', u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]', u'[JavaScript Error: "XML Parsing Error: no root element found\nLocation: https://estilysee.info/\nLine Number 1, Column 1:" {file: "https://estilysee.info/" line: 1}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
estilysee info see bug description url browser version firefox operating system windows tested another browser no problem type something else description site keeps popping up and i don t know what it is steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u from with ❤️
1
690,379
23,656,790,286
IssuesEvent
2022-08-26 12:03:36
iterative/vscode-dvc
https://api.github.com/repos/iterative/vscode-dvc
closed
Timestamp column improvements
enhancement priority-p2 A: table
Related to https://github.com/iterative/vscode-dvc/issues/1444 --- ### EXPs table improvements - [x] Align cell content to the left (Reference https://github.com/iterative/vscode-dvc/issues/1700) - [ ] The column needs to be moveable and hideable https://github.com/iterative/vscode-dvc/issues/2031#issuecomment-1185032012 ### Needs to be fixed - [x] Timestamp label alignment - [x] Cell information alignment ![Screen-Shot-2022-08-09-at-12 06 32-PM](https://user-images.githubusercontent.com/98249521/183706932-8e4f0f63-9c3f-4906-84ff-85353b41ac8e.jpg)
1.0
Timestamp column improvements - Related to https://github.com/iterative/vscode-dvc/issues/1444 --- ### EXPs table improvements - [x] Align cell content to the left (Reference https://github.com/iterative/vscode-dvc/issues/1700) - [ ] The column needs to be moveable and hideable https://github.com/iterative/vscode-dvc/issues/2031#issuecomment-1185032012 ### Needs to be fixed - [x] Timestamp label alignment - [x] Cell information alignment ![Screen-Shot-2022-08-09-at-12 06 32-PM](https://user-images.githubusercontent.com/98249521/183706932-8e4f0f63-9c3f-4906-84ff-85353b41ac8e.jpg)
priority
timestamp column improvements related to exps table improvements align cell content to the left reference the column needs to be moveable and hideable needs to be fixed timestamp label alignment cell information alignment
1
414,631
12,109,366,325
IssuesEvent
2020-04-21 08:38:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
web.skype.com - Browser Unsupported
browser-firefox engine-gecko os-linux priority-important
<!-- @browser: Firefox 75.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:75.0) Gecko/20100101 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/51730 --> **URL**: https://web.skype.com/ **Browser / Version**: Firefox 75.0 **Operating System**: Debian 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: On Firefox I was redirected to https://www.skype.com/en/unsupported-browser/ , stating that I should use Microsoft Edge or Google Chrome. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/0007acc9-e095-4748-ad04-8f12a2aff591.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
web.skype.com - Browser Unsupported - <!-- @browser: Firefox 75.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:75.0) Gecko/20100101 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/51730 --> **URL**: https://web.skype.com/ **Browser / Version**: Firefox 75.0 **Operating System**: Debian 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: On Firefox I was redirected to https://www.skype.com/en/unsupported-browser/ , stating that I should use Microsoft Edge or Google Chrome. <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/0007acc9-e095-4748-ad04-8f12a2aff591.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
web skype com browser unsupported url browser version firefox operating system debian tested another browser yes edge problem type site is not usable description browser unsupported steps to reproduce on firefox i was redirected to stating that i should use microsoft edge or google chrome view the screenshot img alt screenshot src browser configuration none from with ❤️
1
301,239
9,217,981,271
IssuesEvent
2019-03-11 12:14:18
meumobi/ion-employee
https://api.github.com/repos/meumobi/ion-employee
closed
Cordova allow navigation to YouTube urls on iOS
high-priority
need to [allow navigation](https://github.com/apache/cordova-plugin-whitelist) to YouTube urls in our config.xmlfile by adding the below line: `<allow-navigation href="https://*youtube.com/*"/>` [Source](https://www.sebastianschirmer.com/blog/ionic-youtube-embed/) If I tried to embed a youtube video using iframe tag with only this rule `<allow-navigation="://domain.com/">`, it doesn't allow loading of the video in iframe as youtube.com is not listed in allowed domains. [Source](https://issues.apache.org/jira/browse/CB-10709)
1.0
Cordova allow navigation to YouTube urls on iOS - need to [allow navigation](https://github.com/apache/cordova-plugin-whitelist) to YouTube urls in our config.xmlfile by adding the below line: `<allow-navigation href="https://*youtube.com/*"/>` [Source](https://www.sebastianschirmer.com/blog/ionic-youtube-embed/) If I tried to embed a youtube video using iframe tag with only this rule `<allow-navigation="://domain.com/">`, it doesn't allow loading of the video in iframe as youtube.com is not listed in allowed domains. [Source](https://issues.apache.org/jira/browse/CB-10709)
priority
cordova allow navigation to youtube urls on ios need to to youtube urls in our config xmlfile by adding the below line allow navigation href if i tried to embed a youtube video using iframe tag with only this rule it doesn t allow loading of the video in iframe as youtube com is not listed in allowed domains
1
663,172
22,163,362,674
IssuesEvent
2022-06-04 21:07:47
Yoooi0/MultiFunPlayer
https://api.github.com/repos/Yoooi0/MultiFunPlayer
closed
Make smart limit user configurable
enhancement priority-low
* Custom limit slope * Source axis * Preserve current slope settings for R1/R2 * Better explanation of what it does **Sample use case:** > smart limit for twist to not move if stroke axis L0 is pressed against you
1.0
Make smart limit user configurable - * Custom limit slope * Source axis * Preserve current slope settings for R1/R2 * Better explanation of what it does **Sample use case:** > smart limit for twist to not move if stroke axis L0 is pressed against you
priority
make smart limit user configurable custom limit slope source axis preserve current slope settings for better explanation of what it does sample use case smart limit for twist to not move if stroke axis is pressed against you
1
234,884
19,273,899,133
IssuesEvent
2021-12-10 09:34:30
Coderockr/backstage
https://api.github.com/repos/Coderockr/backstage
opened
Symfony UX
frontend component not tested yet template backend tools api language: php
Symfony UX é um conjunto de bibliotecas para facilitar a criação e manipulação de componentes de frontend via PHP usando Turbo e outras bibliotecas. --- https://github.com/symfony/ux https://symfony.com/ux https://github.com/hotwired/turbo
1.0
Symfony UX - Symfony UX é um conjunto de bibliotecas para facilitar a criação e manipulação de componentes de frontend via PHP usando Turbo e outras bibliotecas. --- https://github.com/symfony/ux https://symfony.com/ux https://github.com/hotwired/turbo
non_priority
symfony ux symfony ux é um conjunto de bibliotecas para facilitar a criação e manipulação de componentes de frontend via php usando turbo e outras bibliotecas
0