Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
31,066 | 6,420,959,494 | IssuesEvent | 2017-08-09 02:24:26 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Authcomponent problem if change storage key. | Defect On hold | This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: **3.3.13**.
* Platform and Target: **local**.
### What you did
**I changed the storage key in authcomponent.**
```
$this->loadComponent('Auth', [
'storage' => ['className' => 'Session', 'key' => 'Auth.Admin']
]);
```
### What happened
In a controller that calls any component, the authcomponent does not receive the settings made and the response to the server as unauthorized.
### What you expected to happen
There is permission, so it should not redirect to unauthorized (redirectAction default not configured).
P.S:
I have changed the key of the storage by authcomponent and in a request to a controller that loads another component, analyzing I realized that one of the times "Cake\Controller\Component\AuthComponent::authCheck" is called ( seems to be the last one ) The configured settings are not set and so the key configured in the storage is not found understanding that there is no permission.
I could not understand the relation of a controller any call any component and this problem arises, if there are no extra components in the controller this does not happen.
#### To simulate, do:
* Configure the AuthComponent with a different key in the storage.
* Create a controller and load any component there (initialize function)
* Make the request to this controller already authenticated. You will be redirected to the loginAction default, not the configured._ | 1.0 | Authcomponent problem if change storage key. - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: **3.3.13**.
* Platform and Target: **local**.
### What you did
**I changed the storage key in authcomponent.**
```
$this->loadComponent('Auth', [
'storage' => ['className' => 'Session', 'key' => 'Auth.Admin']
]);
```
### What happened
In a controller that calls any component, the authcomponent does not receive the settings made and the response to the server as unauthorized.
### What you expected to happen
There is permission, so it should not redirect to unauthorized (redirectAction default not configured).
P.S:
I have changed the key of the storage by authcomponent and in a request to a controller that loads another component, analyzing I realized that one of the times "Cake\Controller\Component\AuthComponent::authCheck" is called ( seems to be the last one ) The configured settings are not set and so the key configured in the storage is not found understanding that there is no permission.
I could not understand the relation of a controller any call any component and this problem arises, if there are no extra components in the controller this does not happen.
#### To simulate, do:
* Configure the AuthComponent with a different key in the storage.
* Create a controller and load any component there (initialize function)
* Make the request to this controller already authenticated. You will be redirected to the loginAction default, not the configured._ | defect | authcomponent problem if change storage key this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target local what you did i changed the storage key in authcomponent this loadcomponent auth storage what happened in a controller that calls any component the authcomponent does not receive the settings made and the response to the server as unauthorized what you expected to happen there is permission so it should not redirect to unauthorized redirectaction default not configured p s i have changed the key of the storage by authcomponent and in a request to a controller that loads another component analyzing i realized that one of the times cake controller component authcomponent authcheck is called seems to be the last one the configured settings are not set and so the key configured in the storage is not found understanding that there is no permission i could not understand the relation of a controller any call any component and this problem arises if there are no extra components in the controller this does not happen to simulate do configure the authcomponent with a different key in the storage create a controller and load any component there initialize function make the request to this controller already authenticated you will be redirected to the loginaction default not the configured | 1 |
40,591 | 10,554,281,363 | IssuesEvent | 2019-10-03 19:06:28 | hashicorp/packer | https://api.github.com/repos/hashicorp/packer | closed | Digital Ocean builder with Ansible provisioner. | builder/digitalocean community-supported plugin provisioner/ansible-remote | When filing a bug, please include the following headings if possible. Any
example text in this template can be deleted.
#### Overview of the Issue
Running packer with the digitalocean builder and ansible as a provider. The problem does not see mto be with packer but rather with ansible and being able to connect to the server. From the documentatioon here https://www.packer.io/docs/provisioners/ansible.html
It says packer should create an inventory file - but I do not see any - playbook seems to be using localhost.
I am using a structure/approach that seems to be the convention based on my searches. Not sure if it is a but or if I am doing something wrong?
#### Reproduction Steps
Steps to reproduce this issue
### Packer version
`1.4.3`
### Ansible version
`2.7.5`
### Python version
`3.6.8`
### Simplified Packer Buildfile
```json
{
"variables": {
"do_api_token": "token",
"do_size": "1gb"
},
"builders": [
{
"type": "digitalocean",
"image": "ubuntu-18-04-x64",
"tags": ["ubuntu-18_04", "test"],
"api_token": "{{user `do_api_token`}}",
"name": "ubuntu-18-04-x64_test_{{timestamp}}",
"size": "{{user `do_size`}}",
"region": "nyc1",
"ssh_username": "root"
}
],
"provisioners": [
{
"type": "ansible",
"playbook_file": "./ansible/main.yml"
}
]
}
```
## Ansible playbook
./ansible/main.yml
```yml
---
- hosts: all
tasks:
- name: Update and upgrade apt packages
apt:
upgrade: yes
update_cache: yes
cache_valid_time: 86400 #One day`
```
./ansible/ansible.cfg
```yml
[defaults]
inventory = ./inventory
remote_user = root
`
### Operating system and Environment details
`Distributor ID: Ubuntu
Description: Ubuntu 18.04.3 LTS
Release: 18.04
Codename: bionic
```
OS, Architecture, and any other information you can provide about the
environment.
### Log Fragments and crash.log files
```bash
ubuntu-18-04-x64_test_1569527924 output will be in this color.
==> ubuntu-18-04-x64_test_1569527924: Creating temporary ssh key for droplet...
==> ubuntu-18-04-x64_test_1569527924: Creating droplet...
==> ubuntu-18-04-x64_test_1569527924: Waiting for droplet to become active...
==> ubuntu-18-04-x64_test_1569527924: Using ssh communicator to connect: 165.227.200.189
==> ubuntu-18-04-x64_test_1569527924: Waiting for SSH to become available...
==> ubuntu-18-04-x64_test_1569527924: Connected to SSH!
==> ubuntu-18-04-x64_test_1569527924: Provisioning with Ansible...
==> ubuntu-18-04-x64_test_1569527924: Executing Ansible: ansible-playbook --extra-vars packer_build_name=ubuntu-18-04-x64_mongo-4-2_1569527924 packer_builder_type=digitalocean -o IdentitiesOnly=yes -i /tmp/packer-provisioner-ansible829621104 //home/user/ops/infra-ops/image_builds/packer/mongo-images/ansible/main.yml -e ansible_ssh_private_key_file=/tmp/ansible-key733337365
ubuntu-18-04-x64_test_1569527924:
ubuntu-18-04-x64_test_1569527924: PLAY [all] *********************************************************************
ubuntu-18-04-x64_test_1569527924:
ubuntu-18-04-x64_test_1569527924: TASK [Gathering Facts] *********************************************************
ubuntu-18-04-x64_test_1569527924: RequestsDependencyWarning)
ubuntu-18-04-x64_test_1569527924: fatal: [default]: FAILED! => {"changed": false, "module_stderr": "/bin/sh: 1: /usr/bin/python: not found\nShared connection to 127.0.0.1 closed.\r\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 127}
ubuntu-18-04-x64_test_1569527924: to retry, use: --limit @/home/user/ops/infra-ops/image_builds/packer/mongo-images/ansible/main.retry
ubuntu-18-04-x64_test_1569527924:
ubuntu-18-04-x64_test_1569527924: PLAY RECAP *********************************************************************
ubuntu-18-04-x64_test_1569527924: default : ok=0 changed=0 unreachable=0 failed=1
ubuntu-18-04-x64_test_1569527924:
==> ubuntu-18-04-x64_test_1569527924: Destroying droplet...
==> ubuntu-18-04-x64_test_1569527924: Deleting temporary ssh key...
Build 'ubuntu-18-04-x64_test_1569527924' errored: Error executing Ansible: Non-zero exit status: exit status 2
==> Some builds didn't complete successfully and had errors:
--> ubuntu-18-04-x64_test_1569527924: Error executing Ansible: Non-zero exit status: exit status 2
```
Set the env var `PACKER_LOG=1` for maximum log detail.
| 1.0 | Digital Ocean builder with Ansible provisioner. - When filing a bug, please include the following headings if possible. Any
example text in this template can be deleted.
#### Overview of the Issue
Running packer with the digitalocean builder and ansible as a provider. The problem does not see mto be with packer but rather with ansible and being able to connect to the server. From the documentatioon here https://www.packer.io/docs/provisioners/ansible.html
It says packer should create an inventory file - but I do not see any - playbook seems to be using localhost.
I am using a structure/approach that seems to be the convention based on my searches. Not sure if it is a but or if I am doing something wrong?
#### Reproduction Steps
Steps to reproduce this issue
### Packer version
`1.4.3`
### Ansible version
`2.7.5`
### Python version
`3.6.8`
### Simplified Packer Buildfile
```json
{
"variables": {
"do_api_token": "token",
"do_size": "1gb"
},
"builders": [
{
"type": "digitalocean",
"image": "ubuntu-18-04-x64",
"tags": ["ubuntu-18_04", "test"],
"api_token": "{{user `do_api_token`}}",
"name": "ubuntu-18-04-x64_test_{{timestamp}}",
"size": "{{user `do_size`}}",
"region": "nyc1",
"ssh_username": "root"
}
],
"provisioners": [
{
"type": "ansible",
"playbook_file": "./ansible/main.yml"
}
]
}
```
## Ansible playbook
./ansible/main.yml
```yml
---
- hosts: all
tasks:
- name: Update and upgrade apt packages
apt:
upgrade: yes
update_cache: yes
cache_valid_time: 86400 #One day`
```
./ansible/ansible.cfg
```yml
[defaults]
inventory = ./inventory
remote_user = root
`
### Operating system and Environment details
`Distributor ID: Ubuntu
Description: Ubuntu 18.04.3 LTS
Release: 18.04
Codename: bionic
```
OS, Architecture, and any other information you can provide about the
environment.
### Log Fragments and crash.log files
```bash
ubuntu-18-04-x64_test_1569527924 output will be in this color.
==> ubuntu-18-04-x64_test_1569527924: Creating temporary ssh key for droplet...
==> ubuntu-18-04-x64_test_1569527924: Creating droplet...
==> ubuntu-18-04-x64_test_1569527924: Waiting for droplet to become active...
==> ubuntu-18-04-x64_test_1569527924: Using ssh communicator to connect: 165.227.200.189
==> ubuntu-18-04-x64_test_1569527924: Waiting for SSH to become available...
==> ubuntu-18-04-x64_test_1569527924: Connected to SSH!
==> ubuntu-18-04-x64_test_1569527924: Provisioning with Ansible...
==> ubuntu-18-04-x64_test_1569527924: Executing Ansible: ansible-playbook --extra-vars packer_build_name=ubuntu-18-04-x64_mongo-4-2_1569527924 packer_builder_type=digitalocean -o IdentitiesOnly=yes -i /tmp/packer-provisioner-ansible829621104 //home/user/ops/infra-ops/image_builds/packer/mongo-images/ansible/main.yml -e ansible_ssh_private_key_file=/tmp/ansible-key733337365
ubuntu-18-04-x64_test_1569527924:
ubuntu-18-04-x64_test_1569527924: PLAY [all] *********************************************************************
ubuntu-18-04-x64_test_1569527924:
ubuntu-18-04-x64_test_1569527924: TASK [Gathering Facts] *********************************************************
ubuntu-18-04-x64_test_1569527924: RequestsDependencyWarning)
ubuntu-18-04-x64_test_1569527924: fatal: [default]: FAILED! => {"changed": false, "module_stderr": "/bin/sh: 1: /usr/bin/python: not found\nShared connection to 127.0.0.1 closed.\r\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 127}
ubuntu-18-04-x64_test_1569527924: to retry, use: --limit @/home/user/ops/infra-ops/image_builds/packer/mongo-images/ansible/main.retry
ubuntu-18-04-x64_test_1569527924:
ubuntu-18-04-x64_test_1569527924: PLAY RECAP *********************************************************************
ubuntu-18-04-x64_test_1569527924: default : ok=0 changed=0 unreachable=0 failed=1
ubuntu-18-04-x64_test_1569527924:
==> ubuntu-18-04-x64_test_1569527924: Destroying droplet...
==> ubuntu-18-04-x64_test_1569527924: Deleting temporary ssh key...
Build 'ubuntu-18-04-x64_test_1569527924' errored: Error executing Ansible: Non-zero exit status: exit status 2
==> Some builds didn't complete successfully and had errors:
--> ubuntu-18-04-x64_test_1569527924: Error executing Ansible: Non-zero exit status: exit status 2
```
Set the env var `PACKER_LOG=1` for maximum log detail.
| non_defect | digital ocean builder with ansible provisioner when filing a bug please include the following headings if possible any example text in this template can be deleted overview of the issue running packer with the digitalocean builder and ansible as a provider the problem does not see mto be with packer but rather with ansible and being able to connect to the server from the documentatioon here it says packer should create an inventory file but i do not see any playbook seems to be using localhost i am using a structure approach that seems to be the convention based on my searches not sure if it is a but or if i am doing something wrong reproduction steps steps to reproduce this issue packer version ansible version python version simplified packer buildfile json variables do api token token do size builders type digitalocean image ubuntu tags api token user do api token name ubuntu test timestamp size user do size region ssh username root provisioners type ansible playbook file ansible main yml ansible playbook ansible main yml yml hosts all tasks name update and upgrade apt packages apt upgrade yes update cache yes cache valid time one day ansible ansible cfg yml inventory inventory remote user root operating system and environment details distributor id ubuntu description ubuntu lts release codename bionic os architecture and any other information you can provide about the environment log fragments and crash log files bash ubuntu test output will be in this color ubuntu test creating temporary ssh key for droplet ubuntu test creating droplet ubuntu test waiting for droplet to become active ubuntu test using ssh communicator to connect ubuntu test waiting for ssh to become available ubuntu test connected to ssh ubuntu test provisioning with ansible ubuntu test executing ansible ansible playbook extra vars packer build name ubuntu mongo packer builder type digitalocean o identitiesonly yes i tmp packer provisioner home user ops infra ops image builds packer mongo images ansible main yml e ansible ssh private key file tmp ansible ubuntu test ubuntu test play ubuntu test ubuntu test task ubuntu test requestsdependencywarning ubuntu test fatal failed changed false module stderr bin sh usr bin python not found nshared connection to closed r n module stdout msg module failure nsee stdout stderr for the exact error rc ubuntu test to retry use limit home user ops infra ops image builds packer mongo images ansible main retry ubuntu test ubuntu test play recap ubuntu test default ok changed unreachable failed ubuntu test ubuntu test destroying droplet ubuntu test deleting temporary ssh key build ubuntu test errored error executing ansible non zero exit status exit status some builds didn t complete successfully and had errors ubuntu test error executing ansible non zero exit status exit status set the env var packer log for maximum log detail | 0 |
34,722 | 7,459,421,841 | IssuesEvent | 2018-03-30 15:15:07 | kerdokullamae/test_koik_issued | https://api.github.com/repos/kerdokullamae/test_koik_issued | closed | linkide Eelmine ja Järgmine käitumine | C: AIS P: highest R: fixed T: defect | **Reported by katrin vesterblom on 15 Oct 2015 08:47 UTC**
rahvusarhiiv.tietotest.ee
- teen päringu Pealkiri=Tallinn;
- valin tulemustest esimesel lehel 5-nda "Tallinna vanalinna välisuksed";
- vajutan linki Eelmine, tuleb veateade:
Tekkis viga
Ups, midagi läks valesti!
A form can only be submitted once
Lae leht uuesti või anna teada süsteemi haldajale.
- liigun brauseri Back-nupuga tagasi, vajutan linki Järgmine, tuleb sama veateade.
- teen sama päringu (Pealkiri=Tallinn) uuesti, ja avan taas ka sama KÜ (Sari "..välisuksed");
- valin ühe selle KÜ alluva, Kirjeldushierarhiast;
http://rahvusarhiiv.tietotest.ee/et/description_unit/view/?id=123000003470 .
- vajutan linki Eelmine:
- tuleb miski ERAF.1 fondi Säilik http://rahvusarhiiv.tietotest.ee/et/description_unit/view/?id=111702910115&resultNum=1 .
- ka teistel katsetel kipuvad need lingid peamiselt minema mingitele venekeelsetele säilikutele, mis on (peamiselt?) fondist ERAF.1.
| 1.0 | linkide Eelmine ja Järgmine käitumine - **Reported by katrin vesterblom on 15 Oct 2015 08:47 UTC**
rahvusarhiiv.tietotest.ee
- teen päringu Pealkiri=Tallinn;
- valin tulemustest esimesel lehel 5-nda "Tallinna vanalinna välisuksed";
- vajutan linki Eelmine, tuleb veateade:
Tekkis viga
Ups, midagi läks valesti!
A form can only be submitted once
Lae leht uuesti või anna teada süsteemi haldajale.
- liigun brauseri Back-nupuga tagasi, vajutan linki Järgmine, tuleb sama veateade.
- teen sama päringu (Pealkiri=Tallinn) uuesti, ja avan taas ka sama KÜ (Sari "..välisuksed");
- valin ühe selle KÜ alluva, Kirjeldushierarhiast;
http://rahvusarhiiv.tietotest.ee/et/description_unit/view/?id=123000003470 .
- vajutan linki Eelmine:
- tuleb miski ERAF.1 fondi Säilik http://rahvusarhiiv.tietotest.ee/et/description_unit/view/?id=111702910115&resultNum=1 .
- ka teistel katsetel kipuvad need lingid peamiselt minema mingitele venekeelsetele säilikutele, mis on (peamiselt?) fondist ERAF.1.
| defect | linkide eelmine ja järgmine käitumine reported by katrin vesterblom on oct utc rahvusarhiiv tietotest ee teen päringu pealkiri tallinn valin tulemustest esimesel lehel nda tallinna vanalinna välisuksed vajutan linki eelmine tuleb veateade tekkis viga ups midagi läks valesti a form can only be submitted once lae leht uuesti või anna teada süsteemi haldajale liigun brauseri back nupuga tagasi vajutan linki järgmine tuleb sama veateade teen sama päringu pealkiri tallinn uuesti ja avan taas ka sama kü sari välisuksed valin ühe selle kü alluva kirjeldushierarhiast vajutan linki eelmine tuleb miski eraf fondi säilik ka teistel katsetel kipuvad need lingid peamiselt minema mingitele venekeelsetele säilikutele mis on peamiselt fondist eraf | 1 |
76,746 | 26,575,576,791 | IssuesEvent | 2023-01-21 19:23:57 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | closed | CHANGE_CREATURE_OWNER sometimes fails | Type-Defect Status-Invalid | User provided this script, on a map filled with heroes, it usually fails to convert 5 heroes, more often converts 1~4 heroes. I reproduced it the first try.
```
ADD_OBJECT_TO_LEVEL(SPECBOX_CUSTOM,PLAYER0,1,PLAYER_NEUTRAL)
IF(PLAYER0,BOX1_ACTIVATED >= 1)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,ANYWHERE,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,ANYWHERE,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,ANYWHERE,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,LEAST_EXPERIENCED,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,LEAST_EXPERIENCED,PLAYER0)
ENDIF
``` | 1.0 | CHANGE_CREATURE_OWNER sometimes fails - User provided this script, on a map filled with heroes, it usually fails to convert 5 heroes, more often converts 1~4 heroes. I reproduced it the first try.
```
ADD_OBJECT_TO_LEVEL(SPECBOX_CUSTOM,PLAYER0,1,PLAYER_NEUTRAL)
IF(PLAYER0,BOX1_ACTIVATED >= 1)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,ANYWHERE,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,ANYWHERE,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,ANYWHERE,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,LEAST_EXPERIENCED,PLAYER0)
CHANGE_CREATURE_OWNER(PLAYER_GOOD,ANY_CREATURE,LEAST_EXPERIENCED,PLAYER0)
ENDIF
``` | defect | change creature owner sometimes fails user provided this script on a map filled with heroes it usually fails to convert heroes more often converts heroes i reproduced it the first try add object to level specbox custom player neutral if activated change creature owner player good any creature anywhere change creature owner player good any creature anywhere change creature owner player good any creature anywhere change creature owner player good any creature least experienced change creature owner player good any creature least experienced endif | 1 |
772,214 | 27,111,950,891 | IssuesEvent | 2023-02-15 15:53:15 | tallyhowallet/extension | https://api.github.com/repos/tallyhowallet/extension | closed | Can't LP on Velodrome | Type: Bug Priority: Medium Status: Available | ### Discord Discussion Link
n/a
### What browsers are you seeing the problem on?
Chrome
### What were you trying to do?
Deposit USDC and DAI on Velodrome
### What did not work?
When depositing into an LP on https://app.velodrome.finance/liquidity/create a user can either decide to
1) deposit & stake in a combined flow. Deposit assets, stake LP tokens
2) deposit assets in a separate action, and optionally stake the LP tokens
Both of these consists of multiple transactions, and these aren't handled well.
Pool tested: USDC & DAI (Add liquidity to vAMM-USDC/DAI)
**With flow 2, what should happen:**
- an approve transaction for DAI spend
- an approve transaction for USDC spend
- a deposit transaction
**What happens:**
- two approve transactions for 1 asset (either USDC or DAI) or at least two UI window popovers. If signed, the transaction is dropped and and the UI prompts "the user rejected the request"
- the transaction is not visible on optimistic.etherscan
Dropped tx hash: https://optimistic.etherscan.io/tx/0x7480b9fd2626be7b261c00cb84e36d1c891be4fa7dce6f2811c543b311c0f67a
<img width="1456" alt="Screen Shot 2022-10-03 at 14 20 04" src="https://user-images.githubusercontent.com/71384922/193576034-39dcc603-3568-439c-875d-b38db7ba64d3.png">
<img width="1164" alt="Screen Shot 2022-10-03 at 12 31 07" src="https://user-images.githubusercontent.com/71384922/193576092-b4cf70ba-d408-4acf-9469-7fcba6a5f0f8.png">
Logs:
[logs_v0_16_5__2022-10-03T14_25_55+02_00.txt](https://github.com/tallycash/extension/files/9697434/logs_v0_16_5__2022-10-03T14_25_55%2B02_00.txt)
### Version
v0.16.4
### Relevant log output
_No response_ | 1.0 | Can't LP on Velodrome - ### Discord Discussion Link
n/a
### What browsers are you seeing the problem on?
Chrome
### What were you trying to do?
Deposit USDC and DAI on Velodrome
### What did not work?
When depositing into an LP on https://app.velodrome.finance/liquidity/create a user can either decide to
1) deposit & stake in a combined flow. Deposit assets, stake LP tokens
2) deposit assets in a separate action, and optionally stake the LP tokens
Both of these consists of multiple transactions, and these aren't handled well.
Pool tested: USDC & DAI (Add liquidity to vAMM-USDC/DAI)
**With flow 2, what should happen:**
- an approve transaction for DAI spend
- an approve transaction for USDC spend
- a deposit transaction
**What happens:**
- two approve transactions for 1 asset (either USDC or DAI) or at least two UI window popovers. If signed, the transaction is dropped and and the UI prompts "the user rejected the request"
- the transaction is not visible on optimistic.etherscan
Dropped tx hash: https://optimistic.etherscan.io/tx/0x7480b9fd2626be7b261c00cb84e36d1c891be4fa7dce6f2811c543b311c0f67a
<img width="1456" alt="Screen Shot 2022-10-03 at 14 20 04" src="https://user-images.githubusercontent.com/71384922/193576034-39dcc603-3568-439c-875d-b38db7ba64d3.png">
<img width="1164" alt="Screen Shot 2022-10-03 at 12 31 07" src="https://user-images.githubusercontent.com/71384922/193576092-b4cf70ba-d408-4acf-9469-7fcba6a5f0f8.png">
Logs:
[logs_v0_16_5__2022-10-03T14_25_55+02_00.txt](https://github.com/tallycash/extension/files/9697434/logs_v0_16_5__2022-10-03T14_25_55%2B02_00.txt)
### Version
v0.16.4
### Relevant log output
_No response_ | non_defect | can t lp on velodrome discord discussion link n a what browsers are you seeing the problem on chrome what were you trying to do deposit usdc and dai on velodrome what did not work when depositing into an lp on a user can either decide to deposit stake in a combined flow deposit assets stake lp tokens deposit assets in a separate action and optionally stake the lp tokens both of these consists of multiple transactions and these aren t handled well pool tested usdc dai add liquidity to vamm usdc dai with flow what should happen an approve transaction for dai spend an approve transaction for usdc spend a deposit transaction what happens two approve transactions for asset either usdc or dai or at least two ui window popovers if signed the transaction is dropped and and the ui prompts the user rejected the request the transaction is not visible on optimistic etherscan dropped tx hash img width alt screen shot at src img width alt screen shot at src logs version relevant log output no response | 0 |
18,918 | 3,098,775,471 | IssuesEvent | 2015-08-28 13:22:23 | junichi11/netbeans-backlog-plugin | https://api.github.com/repos/junichi11/netbeans-backlog-plugin | closed | Deadlock when attaching files on the issue pannel | defect | Netbeans stops responding when opening the file dialog to attach files on the issue panel.
Steps to reproduce:
1 - Open an issue.
2 - Click attach files.
Actual results:
Netbeans stops responding and forced termination is required.
Expected results:
See a file chooser dialog, and proceed with the attachment process.
Plugin version:
0.3.2
Netbeans version:
8.0.2 ( OSX 10.9.5 ) | 1.0 | Deadlock when attaching files on the issue pannel - Netbeans stops responding when opening the file dialog to attach files on the issue panel.
Steps to reproduce:
1 - Open an issue.
2 - Click attach files.
Actual results:
Netbeans stops responding and forced termination is required.
Expected results:
See a file chooser dialog, and proceed with the attachment process.
Plugin version:
0.3.2
Netbeans version:
8.0.2 ( OSX 10.9.5 ) | defect | deadlock when attaching files on the issue pannel netbeans stops responding when opening the file dialog to attach files on the issue panel steps to reproduce open an issue click attach files actual results netbeans stops responding and forced termination is required expected results see a file chooser dialog and proceed with the attachment process plugin version netbeans version osx | 1 |
164,242 | 20,364,408,180 | IssuesEvent | 2022-02-21 02:44:06 | dmartinez777/AzureDevOpsAngular | https://api.github.com/repos/dmartinez777/AzureDevOpsAngular | opened | CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2021-27515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.1000.7.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path.
<p>Publish Date: 2021-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p>
<p>Release Date: 2021-02-22</p>
<p>Fix Resolution (url-parse): 1.5.0</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.1000.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz - ## CVE-2021-27515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.1000.7.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path.
<p>Publish Date: 2021-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p>
<p>Release Date: 2021-02-22</p>
<p>Fix Resolution (url-parse): 1.5.0</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.1000.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file package json path to vulnerable library node modules url parse package json dependency hierarchy build angular tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in base branch master vulnerability details url parse before mishandles certain uses of backslash such as http and interprets the uri as a relative path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution angular devkit build angular step up your open source security game with whitesource | 0 |
73,303 | 8,853,610,857 | IssuesEvent | 2019-01-08 21:54:27 | mozilla/foundation.mozilla.org | https://api.github.com/repos/mozilla/foundation.mozilla.org | closed | Auto crop cms images | design | Designers to provide a list of image dimensions for all cms image types so devs can make the cms crop the images.
We should try to be consistent with our ratios as possible so staff can easily use one image in multiple places and not have to crop it to a bunch of different sizes.
### Home
- hero image (desktop & mobile - 4:3 ratio)
- featured news image (desktop & mobile - 16:9 ratio)
- both happening now sections (card desktop & mobile, and wide version - 16:9 ratio)
### Initiatives
- Feature program (desktop & mobile - 16:9 ratio)
### Participate
- hero image - WIP (pending header template)
- small feature cards - 16:9 ratio
- larger features - 16:9 ratio
### Future
- Make sure we do this for upcoming Fellowship and Blog pages.
- Design fallback images for cms images #2416 | 1.0 | Auto crop cms images - Designers to provide a list of image dimensions for all cms image types so devs can make the cms crop the images.
We should try to be consistent with our ratios as possible so staff can easily use one image in multiple places and not have to crop it to a bunch of different sizes.
### Home
- hero image (desktop & mobile - 4:3 ratio)
- featured news image (desktop & mobile - 16:9 ratio)
- both happening now sections (card desktop & mobile, and wide version - 16:9 ratio)
### Initiatives
- Feature program (desktop & mobile - 16:9 ratio)
### Participate
- hero image - WIP (pending header template)
- small feature cards - 16:9 ratio
- larger features - 16:9 ratio
### Future
- Make sure we do this for upcoming Fellowship and Blog pages.
- Design fallback images for cms images #2416 | non_defect | auto crop cms images designers to provide a list of image dimensions for all cms image types so devs can make the cms crop the images we should try to be consistent with our ratios as possible so staff can easily use one image in multiple places and not have to crop it to a bunch of different sizes home hero image desktop mobile ratio featured news image desktop mobile ratio both happening now sections card desktop mobile and wide version ratio initiatives feature program desktop mobile ratio participate hero image wip pending header template small feature cards ratio larger features ratio future make sure we do this for upcoming fellowship and blog pages design fallback images for cms images | 0 |
63,454 | 17,662,425,538 | IssuesEvent | 2021-08-21 19:46:04 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | brentq can overflow / underflow | defect scipy.optimize C/C++ | According to the source code:
https://github.com/scipy/scipy/blob/5f4c4d802e5a56708d86909af6e5685cd95e6e66/scipy/optimize/Zeros/brentq.c#L67
the brentq algorithm uses the product of two function values to see if the zero is in the bracketing interval.
In
Malcolm, Michael A., Cleve B. Moler, and George Elmer Forsythe.
Computer methods for mathematical computations. Prentice-Hall, 1977.
we see the following advice:
"Careful attention is paid to underflow and overflow problems. For example, the test to ensure that F(B) and F(C) have different signs is *not* the conventional:
F(B) * F(C) < 0.0
If F(B)=10^-40 and F(C)=10^-40, then they have opposite signs, but on many computers the product will underflow and be set to zero, and the test will fail."
| 1.0 | brentq can overflow / underflow - According to the source code:
https://github.com/scipy/scipy/blob/5f4c4d802e5a56708d86909af6e5685cd95e6e66/scipy/optimize/Zeros/brentq.c#L67
the brentq algorithm uses the product of two function values to see if the zero is in the bracketing interval.
In
Malcolm, Michael A., Cleve B. Moler, and George Elmer Forsythe.
Computer methods for mathematical computations. Prentice-Hall, 1977.
we see the following advice:
"Careful attention is paid to underflow and overflow problems. For example, the test to ensure that F(B) and F(C) have different signs is *not* the conventional:
F(B) * F(C) < 0.0
If F(B)=10^-40 and F(C)=10^-40, then they have opposite signs, but on many computers the product will underflow and be set to zero, and the test will fail."
| defect | brentq can overflow underflow according to the source code the brentq algorithm uses the product of two function values to see if the zero is in the bracketing interval in malcolm michael a cleve b moler and george elmer forsythe computer methods for mathematical computations prentice hall we see the following advice careful attention is paid to underflow and overflow problems for example the test to ensure that f b and f c have different signs is not the conventional f b f c if f b and f c then they have opposite signs but on many computers the product will underflow and be set to zero and the test will fail | 1 |
69,674 | 22,603,916,759 | IssuesEvent | 2022-06-29 11:38:01 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | zfs send/receive coredump with docker dataset | Type: Defect | This report is for Arch Linux with
```
zfs-2.1.5-1
zfs-kmod-2.1.5-1
```
installed via `zfs-dkms 2.1.5-1`
kernel is either 5.17.8 or 5.15.50. It happens with both.
I experience crashes when trying to send/receive a docker dataset. I am using send/receive for my regular backus with many other datasets and never experienced an issue like this. It seems to be related to this docker dataset.
If I repeat the zfs send/receive command often enough it finally succeeds at some point in time. But only after multiple tries and multiple coredumps.
```
# zfs send -I 'zstore/docker'@'2022-04-24--09:37-zf1' 'zstore/docker'@'2022-06-29--08:48-zf1' | zfs receive -s -F 'zstore/docker-test'
cannot receive: failed to read from stream
zsh: segmentation fault (core dumped) zfs send -I 'zstore/docker'@'2022-04-24--09:37-zf1' |
zsh: exit 1 zfs receive -s -F 'zstore/docker-test'
```
coredump info:
```
coredumpctl info
PID: 34194 (zfs)
UID: 0 (root)
GID: 0 (root)
Signal: 11 (SEGV)
Timestamp: Wed 2022-06-29 13:28:15 CEST (25s ago)
Command Line: zfs send -I zstore/docker@2022-04-24--09:37-zf1 zstore/docker@2022-06-29--08:48-zf1
Executable: /usr/bin/zfs
Control Group: /system.slice/sshd.service
Unit: sshd.service
Slice: system.slice
Boot ID: e3c510ea89b94931beed469f23be9947
Machine ID: 4bd88beaa35549b5922de02c8064cbf1
Hostname: rakete
Storage: /var/lib/systemd/coredump/core.zfs.0.e3c510ea89b94931beed469f23be9947.34194.1656502095000000.zst (inaccessibl>
Message: Process 34194 (zfs) of user 0 dumped core.
Module linux-vdso.so.1 with build-id f8a28135883cc0ea0e8b29412015dca150c6108b
Module libresolv.so.2 with build-id 89a368a6ad1b392d126a2a5beb9c2f61ade00279
Module libkeyutils.so.1 with build-id ac405ddd17be10ce538da3211415ee50c8f8df79
Module libkrb5support.so.0 with build-id 15f223925ef59dee4379ebbc0fcd14eda9ba81a2
Module libcom_err.so.2 with build-id 3360a28740ffbbd5a5c0c21d09072445908707e5
Module libk5crypto.so.3 with build-id cc77a742cb62447a53d98285b41558b8acd92866
Module libkrb5.so.3 with build-id 371cc767dacb17cb42c9c44b88eebbed5ee9a756
Module libpthread.so.0 with build-id 95ae4f30a6f12ccbff645d30f8e1a3ee23ec7d36
Module libgssapi_krb5.so.2 with build-id 292f1ce32161c0ecc4a287bc8494d5d7c420a03f
Module ld-linux-x86-64.so.2 with build-id 0effd0e43efa4468d3c31871c93af0b7f3005673
Module libgcc_s.so.1 with build-id 0e3de903950e35ae59a5de8c00b1817a4a71ca01
Module libz.so.1 with build-id fefe3219a96d682ec98fcfb78866b8594298b5a2
Module libcrypto.so.1.1 with build-id d1f36af479cd3316f5ea2460b330fbe703587f12
Module libm.so.6 with build-id 1b7296ef9fd806e47060788389293c824b09ad72
Module libtirpc.so.3 with build-id 5bef2adfdee3df283f593b3e2d37b6dac405256a
Module libudev.so.1 with build-id 541e6841430a5ee36134325ec0ce669c2c0b9053
Module libblkid.so.1 with build-id 140694a62d8d4d07c6c320a501f948dd1b389d73
Module libuuid.so.1 with build-id 032a21acd159ee3902605e9911be5f86a7df7df9
Module libc.so.6 with build-id 60df1df31f02a7b23da83e8ef923359885b81492
Module libuutil.so.3 with build-id 79a31f3c024a9e7da5e71c781f9017a9e2b229d5
Module libnvpair.so.3 with build-id 9907f66528dacfcf4e3d2ccdcf2d64a4cb07c158
Module libzfs_core.so.3 with build-id 559a5214d79feaad5eca9dfc013170effd2acea4
Module libzfs.so.4 with build-id 935f3f20dd39c007f5c24ff27bae869c7f37163d
Module zfs with build-id 72125cf8e3782c4f34af18dc010ed1d99eb7a087
Stack trace of thread 34194:
#0 0x00007f51c664ddd4 fletcher_4_incremental_native (libzfs.so.4 + 0x4bdd4)
#1 0x00007f51c66379d2 n/a (libzfs.so.4 + 0x359d2)
#2 0x00007f51c663d7a7 zfs_send (libzfs.so.4 + 0x3b7a7)
#3 0x0000563f2f7d2194 n/a (zfs + 0xe194)
#4 0x0000563f2f7ca364 n/a (zfs + 0x6364)
#5 0x00007f51c63d2290 n/a (libc.so.6 + 0x29290)
#6 0x00007f51c63d234a __libc_start_main (libc.so.6 + 0x2934a)
#7 0x0000563f2f7ca485 n/a (zfs + 0x6485)
ELF object binary architecture: AMD x86-64
```
```
# zfs get all zstore/docker
NAME PROPERTY VALUE SOURCE
zstore/docker type filesystem -
zstore/docker creation Fr Apr 1 18:42 2022 -
zstore/docker used 1.07G -
zstore/docker available 1.83T -
zstore/docker referenced 113M -
zstore/docker compressratio 1.72x -
zstore/docker mounted yes -
zstore/docker quota none default
zstore/docker reservation none default
zstore/docker recordsize 128K local
zstore/docker mountpoint /var/lib/docker local
zstore/docker sharenfs off default
zstore/docker checksum on default
zstore/docker compression lz4 inherited from zstore
zstore/docker atime on inherited from zstore
zstore/docker devices on default
zstore/docker exec on default
zstore/docker setuid on default
zstore/docker readonly off inherited from zstore
zstore/docker zoned off default
zstore/docker snapdir hidden default
zstore/docker aclmode discard default
zstore/docker aclinherit restricted default
zstore/docker createtxg 1620621 -
zstore/docker canmount on default
zstore/docker xattr sa inherited from zstore
zstore/docker copies 1 default
zstore/docker version 5 -
zstore/docker utf8only off -
zstore/docker normalization none -
zstore/docker casesensitivity sensitive -
zstore/docker vscan off default
zstore/docker nbmand off default
zstore/docker sharesmb off default
zstore/docker refquota none default
zstore/docker refreservation none default
zstore/docker guid 1018209730648405787 -
zstore/docker primarycache all default
zstore/docker secondarycache all default
zstore/docker usedbysnapshots 224M -
zstore/docker usedbydataset 113M -
zstore/docker usedbychildren 763M -
zstore/docker usedbyrefreservation 0B -
zstore/docker logbias latency default
zstore/docker objsetid 42266 -
zstore/docker dedup off default
zstore/docker mlslabel none default
zstore/docker sync standard default
zstore/docker dnodesize legacy default
zstore/docker refcompressratio 1.33x -
zstore/docker written 2.87M -
zstore/docker logicalused 1.71G -
zstore/docker logicalreferenced 149M -
zstore/docker volmode default default
zstore/docker filesystem_limit none default
zstore/docker snapshot_limit none default
zstore/docker filesystem_count none default
zstore/docker snapshot_count none default
zstore/docker snapdev hidden default
zstore/docker acltype posix inherited from zstore
zstore/docker context none default
zstore/docker fscontext none default
zstore/docker defcontext none default
zstore/docker rootcontext none default
zstore/docker relatime on inherited from zstore
zstore/docker redundant_metadata all default
zstore/docker overlay on default
zstore/docker encryption off default
zstore/docker keylocation none default
zstore/docker keyformat none default
zstore/docker pbkdf2iters 0 default
zstore/docker special_small_blocks 0 default
``` | 1.0 | zfs send/receive coredump with docker dataset - This report is for Arch Linux with
```
zfs-2.1.5-1
zfs-kmod-2.1.5-1
```
installed via `zfs-dkms 2.1.5-1`
kernel is either 5.17.8 or 5.15.50. It happens with both.
I experience crashes when trying to send/receive a docker dataset. I am using send/receive for my regular backus with many other datasets and never experienced an issue like this. It seems to be related to this docker dataset.
If I repeat the zfs send/receive command often enough it finally succeeds at some point in time. But only after multiple tries and multiple coredumps.
```
# zfs send -I 'zstore/docker'@'2022-04-24--09:37-zf1' 'zstore/docker'@'2022-06-29--08:48-zf1' | zfs receive -s -F 'zstore/docker-test'
cannot receive: failed to read from stream
zsh: segmentation fault (core dumped) zfs send -I 'zstore/docker'@'2022-04-24--09:37-zf1' |
zsh: exit 1 zfs receive -s -F 'zstore/docker-test'
```
coredump info:
```
coredumpctl info
PID: 34194 (zfs)
UID: 0 (root)
GID: 0 (root)
Signal: 11 (SEGV)
Timestamp: Wed 2022-06-29 13:28:15 CEST (25s ago)
Command Line: zfs send -I zstore/docker@2022-04-24--09:37-zf1 zstore/docker@2022-06-29--08:48-zf1
Executable: /usr/bin/zfs
Control Group: /system.slice/sshd.service
Unit: sshd.service
Slice: system.slice
Boot ID: e3c510ea89b94931beed469f23be9947
Machine ID: 4bd88beaa35549b5922de02c8064cbf1
Hostname: rakete
Storage: /var/lib/systemd/coredump/core.zfs.0.e3c510ea89b94931beed469f23be9947.34194.1656502095000000.zst (inaccessibl>
Message: Process 34194 (zfs) of user 0 dumped core.
Module linux-vdso.so.1 with build-id f8a28135883cc0ea0e8b29412015dca150c6108b
Module libresolv.so.2 with build-id 89a368a6ad1b392d126a2a5beb9c2f61ade00279
Module libkeyutils.so.1 with build-id ac405ddd17be10ce538da3211415ee50c8f8df79
Module libkrb5support.so.0 with build-id 15f223925ef59dee4379ebbc0fcd14eda9ba81a2
Module libcom_err.so.2 with build-id 3360a28740ffbbd5a5c0c21d09072445908707e5
Module libk5crypto.so.3 with build-id cc77a742cb62447a53d98285b41558b8acd92866
Module libkrb5.so.3 with build-id 371cc767dacb17cb42c9c44b88eebbed5ee9a756
Module libpthread.so.0 with build-id 95ae4f30a6f12ccbff645d30f8e1a3ee23ec7d36
Module libgssapi_krb5.so.2 with build-id 292f1ce32161c0ecc4a287bc8494d5d7c420a03f
Module ld-linux-x86-64.so.2 with build-id 0effd0e43efa4468d3c31871c93af0b7f3005673
Module libgcc_s.so.1 with build-id 0e3de903950e35ae59a5de8c00b1817a4a71ca01
Module libz.so.1 with build-id fefe3219a96d682ec98fcfb78866b8594298b5a2
Module libcrypto.so.1.1 with build-id d1f36af479cd3316f5ea2460b330fbe703587f12
Module libm.so.6 with build-id 1b7296ef9fd806e47060788389293c824b09ad72
Module libtirpc.so.3 with build-id 5bef2adfdee3df283f593b3e2d37b6dac405256a
Module libudev.so.1 with build-id 541e6841430a5ee36134325ec0ce669c2c0b9053
Module libblkid.so.1 with build-id 140694a62d8d4d07c6c320a501f948dd1b389d73
Module libuuid.so.1 with build-id 032a21acd159ee3902605e9911be5f86a7df7df9
Module libc.so.6 with build-id 60df1df31f02a7b23da83e8ef923359885b81492
Module libuutil.so.3 with build-id 79a31f3c024a9e7da5e71c781f9017a9e2b229d5
Module libnvpair.so.3 with build-id 9907f66528dacfcf4e3d2ccdcf2d64a4cb07c158
Module libzfs_core.so.3 with build-id 559a5214d79feaad5eca9dfc013170effd2acea4
Module libzfs.so.4 with build-id 935f3f20dd39c007f5c24ff27bae869c7f37163d
Module zfs with build-id 72125cf8e3782c4f34af18dc010ed1d99eb7a087
Stack trace of thread 34194:
#0 0x00007f51c664ddd4 fletcher_4_incremental_native (libzfs.so.4 + 0x4bdd4)
#1 0x00007f51c66379d2 n/a (libzfs.so.4 + 0x359d2)
#2 0x00007f51c663d7a7 zfs_send (libzfs.so.4 + 0x3b7a7)
#3 0x0000563f2f7d2194 n/a (zfs + 0xe194)
#4 0x0000563f2f7ca364 n/a (zfs + 0x6364)
#5 0x00007f51c63d2290 n/a (libc.so.6 + 0x29290)
#6 0x00007f51c63d234a __libc_start_main (libc.so.6 + 0x2934a)
#7 0x0000563f2f7ca485 n/a (zfs + 0x6485)
ELF object binary architecture: AMD x86-64
```
```
# zfs get all zstore/docker
NAME PROPERTY VALUE SOURCE
zstore/docker type filesystem -
zstore/docker creation Fr Apr 1 18:42 2022 -
zstore/docker used 1.07G -
zstore/docker available 1.83T -
zstore/docker referenced 113M -
zstore/docker compressratio 1.72x -
zstore/docker mounted yes -
zstore/docker quota none default
zstore/docker reservation none default
zstore/docker recordsize 128K local
zstore/docker mountpoint /var/lib/docker local
zstore/docker sharenfs off default
zstore/docker checksum on default
zstore/docker compression lz4 inherited from zstore
zstore/docker atime on inherited from zstore
zstore/docker devices on default
zstore/docker exec on default
zstore/docker setuid on default
zstore/docker readonly off inherited from zstore
zstore/docker zoned off default
zstore/docker snapdir hidden default
zstore/docker aclmode discard default
zstore/docker aclinherit restricted default
zstore/docker createtxg 1620621 -
zstore/docker canmount on default
zstore/docker xattr sa inherited from zstore
zstore/docker copies 1 default
zstore/docker version 5 -
zstore/docker utf8only off -
zstore/docker normalization none -
zstore/docker casesensitivity sensitive -
zstore/docker vscan off default
zstore/docker nbmand off default
zstore/docker sharesmb off default
zstore/docker refquota none default
zstore/docker refreservation none default
zstore/docker guid 1018209730648405787 -
zstore/docker primarycache all default
zstore/docker secondarycache all default
zstore/docker usedbysnapshots 224M -
zstore/docker usedbydataset 113M -
zstore/docker usedbychildren 763M -
zstore/docker usedbyrefreservation 0B -
zstore/docker logbias latency default
zstore/docker objsetid 42266 -
zstore/docker dedup off default
zstore/docker mlslabel none default
zstore/docker sync standard default
zstore/docker dnodesize legacy default
zstore/docker refcompressratio 1.33x -
zstore/docker written 2.87M -
zstore/docker logicalused 1.71G -
zstore/docker logicalreferenced 149M -
zstore/docker volmode default default
zstore/docker filesystem_limit none default
zstore/docker snapshot_limit none default
zstore/docker filesystem_count none default
zstore/docker snapshot_count none default
zstore/docker snapdev hidden default
zstore/docker acltype posix inherited from zstore
zstore/docker context none default
zstore/docker fscontext none default
zstore/docker defcontext none default
zstore/docker rootcontext none default
zstore/docker relatime on inherited from zstore
zstore/docker redundant_metadata all default
zstore/docker overlay on default
zstore/docker encryption off default
zstore/docker keylocation none default
zstore/docker keyformat none default
zstore/docker pbkdf2iters 0 default
zstore/docker special_small_blocks 0 default
``` | defect | zfs send receive coredump with docker dataset this report is for arch linux with zfs zfs kmod installed via zfs dkms kernel is either or it happens with both i experience crashes when trying to send receive a docker dataset i am using send receive for my regular backus with many other datasets and never experienced an issue like this it seems to be related to this docker dataset if i repeat the zfs send receive command often enough it finally succeeds at some point in time but only after multiple tries and multiple coredumps zfs send i zstore docker zstore docker zfs receive s f zstore docker test cannot receive failed to read from stream zsh segmentation fault core dumped zfs send i zstore docker zsh exit zfs receive s f zstore docker test coredump info coredumpctl info pid zfs uid root gid root signal segv timestamp wed cest ago command line zfs send i zstore docker zstore docker executable usr bin zfs control group system slice sshd service unit sshd service slice system slice boot id machine id hostname rakete storage var lib systemd coredump core zfs zst inaccessibl message process zfs of user dumped core module linux vdso so with build id module libresolv so with build id module libkeyutils so with build id module so with build id module libcom err so with build id module so with build id module so with build id module libpthread so with build id module libgssapi so with build id module ld linux so with build id module libgcc s so with build id module libz so with build id module libcrypto so with build id module libm so with build id module libtirpc so with build id module libudev so with build id module libblkid so with build id module libuuid so with build id module libc so with build id module libuutil so with build id module libnvpair so with build id module libzfs core so with build id module libzfs so with build id module zfs with build id stack trace of thread fletcher incremental native libzfs so n a libzfs so zfs send libzfs so n a zfs n a zfs n a libc so libc start main libc so n a zfs elf object binary architecture amd zfs get all zstore docker name property value source zstore docker type filesystem zstore docker creation fr apr zstore docker used zstore docker available zstore docker referenced zstore docker compressratio zstore docker mounted yes zstore docker quota none default zstore docker reservation none default zstore docker recordsize local zstore docker mountpoint var lib docker local zstore docker sharenfs off default zstore docker checksum on default zstore docker compression inherited from zstore zstore docker atime on inherited from zstore zstore docker devices on default zstore docker exec on default zstore docker setuid on default zstore docker readonly off inherited from zstore zstore docker zoned off default zstore docker snapdir hidden default zstore docker aclmode discard default zstore docker aclinherit restricted default zstore docker createtxg zstore docker canmount on default zstore docker xattr sa inherited from zstore zstore docker copies default zstore docker version zstore docker off zstore docker normalization none zstore docker casesensitivity sensitive zstore docker vscan off default zstore docker nbmand off default zstore docker sharesmb off default zstore docker refquota none default zstore docker refreservation none default zstore docker guid zstore docker primarycache all default zstore docker secondarycache all default zstore docker usedbysnapshots zstore docker usedbydataset zstore docker usedbychildren zstore docker usedbyrefreservation zstore docker logbias latency default zstore docker objsetid zstore docker dedup off default zstore docker mlslabel none default zstore docker sync standard default zstore docker dnodesize legacy default zstore docker refcompressratio zstore docker written zstore docker logicalused zstore docker logicalreferenced zstore docker volmode default default zstore docker filesystem limit none default zstore docker snapshot limit none default zstore docker filesystem count none default zstore docker snapshot count none default zstore docker snapdev hidden default zstore docker acltype posix inherited from zstore zstore docker context none default zstore docker fscontext none default zstore docker defcontext none default zstore docker rootcontext none default zstore docker relatime on inherited from zstore zstore docker redundant metadata all default zstore docker overlay on default zstore docker encryption off default zstore docker keylocation none default zstore docker keyformat none default zstore docker default zstore docker special small blocks default | 1 |
47,310 | 5,886,092,109 | IssuesEvent | 2017-05-17 01:04:31 | broadinstitute/gatk | https://api.github.com/repos/broadinstitute/gatk | closed | PSUtilsTest should be rewritten so it succeeds when it runs as root | bug tests | PSUtilsTest should be rewritten so it succeeds when it runs as root. It could possibly make use of `BaseTest.getSafeNonExistentFile()`. | 1.0 | PSUtilsTest should be rewritten so it succeeds when it runs as root - PSUtilsTest should be rewritten so it succeeds when it runs as root. It could possibly make use of `BaseTest.getSafeNonExistentFile()`. | non_defect | psutilstest should be rewritten so it succeeds when it runs as root psutilstest should be rewritten so it succeeds when it runs as root it could possibly make use of basetest getsafenonexistentfile | 0 |
1,398 | 2,603,846,650 | IssuesEvent | 2015-02-24 18:16:07 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳尖锐性疣 | auto-migrated Priority-Medium Type-Defect | ```
沈阳尖锐性疣〓沈陽軍區政治部醫院性病〓TEL:024-31023308〓��
�立于1946年,68年專注于性傳播疾病的研究和治療。位于沈陽�
��沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史悠�
��、設備精良、技術權威、專家云集,是預防、保健、醫療、
科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫��
�、全國首批醫療規范定點單位,是第四軍醫大學、東南大學�
��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部
衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:12 | 1.0 | 沈阳尖锐性疣 - ```
沈阳尖锐性疣〓沈陽軍區政治部醫院性病〓TEL:024-31023308〓��
�立于1946年,68年專注于性傳播疾病的研究和治療。位于沈陽�
��沈河區二緯路32號。是一所與新中國同建立共輝煌的歷史悠�
��、設備精良、技術權威、專家云集,是預防、保健、醫療、
科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫��
�、全國首批醫療規范定點單位,是第四軍醫大學、東南大學�
��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部
衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:12 | defect | 沈阳尖锐性疣 沈阳尖锐性疣〓沈陽軍區政治部醫院性病〓tel: 〓�� � , 。位于沈陽� �� 。是一所與新中國同建立共輝煌的歷史悠� ��、設備精良、技術權威、專家云集,是預防、保健、醫療、 科研康復為一體的綜合性醫院。是國家首批公立甲等部隊醫�� �、全國首批醫療規范定點單位,是第四軍醫大學、東南大學� ��知名高等院校的教學醫院。曾被中國人民解放軍空軍后勤部 衛生部評為衛生工作先進單位,先后兩次榮立集體二等功。 original issue reported on code google com by gmail com on jun at | 1 |
16,356 | 2,889,793,406 | IssuesEvent | 2015-06-13 19:20:36 | damonkohler/sl4a | https://api.github.com/repos/damonkohler/sl4a | opened | Samsung Galaxy (i7500) hang problem | auto-migrated Priority-Medium Type-Defect | _From @GoogleCodeExporter on May 31, 2015 11:27_
```
sl4a installed on a samsung Galaxy (i7500)
Running 1.6 Galaxo V 1.6.3.4
Also At least the last version of ASE hangs with any 1.5 or 1.6 ROM
with the samsung Galaxy, tested that.
What steps will reproduce the problem?
start sl4a | view interpreters | python | exit it all by pressing "back button"
| press "lock screen" ---> and all buttons and display are completely non
responding. It seems not to hang everytime, still ive managed to collect from
logcat and traces from a hang.
I am expecting a clean exit from the sl4a that doesnt make the UI unresponsive
on the telephone.
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `pelle.ot...@gmail.com` on 11 Sep 2010 at 12:10
Attachments:
* [traces.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-427/comment-0/traces.txt)
* [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-427/comment-0/logcat.txt)
_Copied from original issue: damonkohler/android-scripting#427_ | 1.0 | Samsung Galaxy (i7500) hang problem - _From @GoogleCodeExporter on May 31, 2015 11:27_
```
sl4a installed on a samsung Galaxy (i7500)
Running 1.6 Galaxo V 1.6.3.4
Also At least the last version of ASE hangs with any 1.5 or 1.6 ROM
with the samsung Galaxy, tested that.
What steps will reproduce the problem?
start sl4a | view interpreters | python | exit it all by pressing "back button"
| press "lock screen" ---> and all buttons and display are completely non
responding. It seems not to hang everytime, still ive managed to collect from
logcat and traces from a hang.
I am expecting a clean exit from the sl4a that doesnt make the UI unresponsive
on the telephone.
What version of the product are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `pelle.ot...@gmail.com` on 11 Sep 2010 at 12:10
Attachments:
* [traces.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-427/comment-0/traces.txt)
* [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-427/comment-0/logcat.txt)
_Copied from original issue: damonkohler/android-scripting#427_ | defect | samsung galaxy hang problem from googlecodeexporter on may installed on a samsung galaxy running galaxo v also at least the last version of ase hangs with any or rom with the samsung galaxy tested that what steps will reproduce the problem start view interpreters python exit it all by pressing back button press lock screen and all buttons and display are completely non responding it seems not to hang everytime still ive managed to collect from logcat and traces from a hang i am expecting a clean exit from the that doesnt make the ui unresponsive on the telephone what version of the product are you using on what operating system please provide any additional information below original issue reported on code google com by pelle ot gmail com on sep at attachments copied from original issue damonkohler android scripting | 1 |
273,394 | 29,820,295,337 | IssuesEvent | 2023-06-17 01:22:59 | pazhanivel07/frameworks_base_2021-0970 | https://api.github.com/repos/pazhanivel07/frameworks_base_2021-0970 | closed | CVE-2020-0045 (Medium) detected in baseandroid-10.0.0_r44 - autoclosed | Mend: dependency security vulnerability | ## CVE-2020-0045 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary>
<p>
<p>Android framework classes and services</p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_2021-0970/commit/ad3ed522c9ac4d72ed6d51ed523780ad73330cbe">ad3ed522c9ac4d72ed6d51ed523780ad73330cbe</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/cmds/statsd/src/StatsService.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In StatsService::command of StatsService.cpp, there is possible memory corruption due to a race condition. This could lead to local escalation of privilege with System execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-141243101
<p>Publish Date: 2020-03-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0045>CVE-2020-0045</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-0045">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-0045</a></p>
<p>Release Date: 2020-03-10</p>
<p>Fix Resolution: android-10.0.0_r30</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-0045 (Medium) detected in baseandroid-10.0.0_r44 - autoclosed - ## CVE-2020-0045 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary>
<p>
<p>Android framework classes and services</p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_2021-0970/commit/ad3ed522c9ac4d72ed6d51ed523780ad73330cbe">ad3ed522c9ac4d72ed6d51ed523780ad73330cbe</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/cmds/statsd/src/StatsService.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In StatsService::command of StatsService.cpp, there is possible memory corruption due to a race condition. This could lead to local escalation of privilege with System execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-141243101
<p>Publish Date: 2020-03-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0045>CVE-2020-0045</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-0045">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-0045</a></p>
<p>Release Date: 2020-03-10</p>
<p>Fix Resolution: android-10.0.0_r30</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in baseandroid autoclosed cve medium severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files cmds statsd src statsservice cpp vulnerability details in statsservice command of statsservice cpp there is possible memory corruption due to a race condition this could lead to local escalation of privilege with system execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend | 0 |
3,527 | 2,610,064,270 | IssuesEvent | 2015-02-26 18:18:54 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 黄岩检查不育哪家效果最好 | auto-migrated Priority-Medium Type-Defect | ```
黄岩检查不育哪家效果最好【台州五洲生殖医院】24小时健康
咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台
州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1
08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、
109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:04 | 1.0 | 黄岩检查不育哪家效果最好 - ```
黄岩检查不育哪家效果最好【台州五洲生殖医院】24小时健康
咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台
州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、1
08、118、198及椒江一金清公交车直达枫南小区,乘坐107、105、
109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:04 | defect | 黄岩检查不育哪家效果最好 黄岩检查不育哪家效果最好【台州五洲生殖医院】 咨询热线 微信号tzwzszyy 医院地址 台 (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
18,983 | 10,277,577,635 | IssuesEvent | 2019-08-25 06:17:27 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | `tf.keras.layers.Embedding` causes memory leak | 2.0.0-beta0 comp:keras type:bug type:performance | **System information**
- Have I written custom code: yes
- OS Platform and Distribution: Linux Mint 19.1
- TensorFlow installed from: binary (using pip)
- TensorFlow version: 2.0.0-beta1 (v2.0.0-beta0-16-g1d91213fe7)
- Python version: 3.6.8
- CUDA/cuDNN version: 10.0 / 7.5
- GPU model and memory: Nvidia Quadro P1000 - 4 GB GDDR5
**Describe the current behavior**
A GPU (_edit: CPU as well, see addendum below_) memory leak (rapidly) emerges from using (high-dimensional) `tf.keras.layers.Embedding` layers.
To be more precise, I am working on Transformer networks, and found out that when I try to fit one, _e.g._ on the portuguese-to-english translation task presented in [this official tutorial](https://www.tensorflow.org/beta/tutorials/text/transformer), a GPU memory leak emerges after a few iterations. Based on [this StackOverflow post](https://stackoverflow.com/questions/42499592/resourceexhaustederror-oom-when-allocating-tensor-with-shape#42512916), I rapidly came to suspect that the issue comes from the (learnable) embedding layers at the base of both the encoder and decoder parts of the network.
To further assess the issue and its source, I implemented a pseudo-Transformer network (see code linked below) that is stripped of most technical components the actual model embarks (_e.g._ I removed positional encoding, residual connections, masking mechanisms, etc.) - the rationale being to provide a more condense (and faster-run) code to document this issue, but also to confirm that the leak does not come from custom layers or any "complex" data processing mechanism.
The provided code includes a data pre-processing pipeline entirely based on the aforementioned [tutorial](https://www.tensorflow.org/beta/tutorials/text/transformer), a model-construction function that makes use of the keras functional API, and a main function to call the former and start the fitting process. On my computer, everything runs fine and I can see the first few fitting iterations pass, until an ugly stack of allocation error messages show up (see full log linked below), whose informative part seems to be `W tensorflow/core/framework/op_kernel.cc:1546] OP_REQUIRES failed at cwise_ops_common.cc:70 : Resource exhausted: OOM when allocating tensor`
**Addendum**: I re-ran the provided code disabling access to the GPU, and it turns out there also is a high memory usage when running on CPU. During the first epoch (and mostly during its first half), memory usage goes up multiple GB (in my case, from 2 to 10 GB, with an increase from 2 to 7 within the first 60 train steps out of 704), and keeps slowly increasing throughout the following epochs (with minor decreases between increases, thus displaying local plateaux which I would guess are related to the loading / discarding of data batches). Although it is a bit less of a problem than with GPU since it is relatively common to have quite some RAM available (plus some swap space, on linux), it still does not feel right that fitting the fake model on a dataset which can be fully loaded in memory (creating a list of Eager Tensors from the `tf.data.Dataset` object containing the batched, padded training set results in a marginal usage of around 100 MB of RAM) would end up using 16GB or RAM. I would also like to note that calling `gc.collect` after training _does not_ empty the used RAM, which is only freed (instantly) when ending the python process.
**Describe the expected behavior**
The fitting process should go one fine, and the memory should not get saturated (I would expect some tensors to be de-allocated as iterations pass).
**Code to reproduce the issue**
The script I wrote to illustrate the issue is publicly accessible as a gist [here](https://gist.github.com/pandrey-fr/c4ba8022c5dd956388e984f49c89ce61).
**Other info / logs**
The full error stack (_with GPU enabled_) is publicly accessible as a gist [here](https://gist.github.com/pandrey-fr/ff004b4cdd6d22b9cd84f82ef4e3a5ac) | True | `tf.keras.layers.Embedding` causes memory leak - **System information**
- Have I written custom code: yes
- OS Platform and Distribution: Linux Mint 19.1
- TensorFlow installed from: binary (using pip)
- TensorFlow version: 2.0.0-beta1 (v2.0.0-beta0-16-g1d91213fe7)
- Python version: 3.6.8
- CUDA/cuDNN version: 10.0 / 7.5
- GPU model and memory: Nvidia Quadro P1000 - 4 GB GDDR5
**Describe the current behavior**
A GPU (_edit: CPU as well, see addendum below_) memory leak (rapidly) emerges from using (high-dimensional) `tf.keras.layers.Embedding` layers.
To be more precise, I am working on Transformer networks, and found out that when I try to fit one, _e.g._ on the portuguese-to-english translation task presented in [this official tutorial](https://www.tensorflow.org/beta/tutorials/text/transformer), a GPU memory leak emerges after a few iterations. Based on [this StackOverflow post](https://stackoverflow.com/questions/42499592/resourceexhaustederror-oom-when-allocating-tensor-with-shape#42512916), I rapidly came to suspect that the issue comes from the (learnable) embedding layers at the base of both the encoder and decoder parts of the network.
To further assess the issue and its source, I implemented a pseudo-Transformer network (see code linked below) that is stripped of most technical components the actual model embarks (_e.g._ I removed positional encoding, residual connections, masking mechanisms, etc.) - the rationale being to provide a more condense (and faster-run) code to document this issue, but also to confirm that the leak does not come from custom layers or any "complex" data processing mechanism.
The provided code includes a data pre-processing pipeline entirely based on the aforementioned [tutorial](https://www.tensorflow.org/beta/tutorials/text/transformer), a model-construction function that makes use of the keras functional API, and a main function to call the former and start the fitting process. On my computer, everything runs fine and I can see the first few fitting iterations pass, until an ugly stack of allocation error messages show up (see full log linked below), whose informative part seems to be `W tensorflow/core/framework/op_kernel.cc:1546] OP_REQUIRES failed at cwise_ops_common.cc:70 : Resource exhausted: OOM when allocating tensor`
**Addendum**: I re-ran the provided code disabling access to the GPU, and it turns out there also is a high memory usage when running on CPU. During the first epoch (and mostly during its first half), memory usage goes up multiple GB (in my case, from 2 to 10 GB, with an increase from 2 to 7 within the first 60 train steps out of 704), and keeps slowly increasing throughout the following epochs (with minor decreases between increases, thus displaying local plateaux which I would guess are related to the loading / discarding of data batches). Although it is a bit less of a problem than with GPU since it is relatively common to have quite some RAM available (plus some swap space, on linux), it still does not feel right that fitting the fake model on a dataset which can be fully loaded in memory (creating a list of Eager Tensors from the `tf.data.Dataset` object containing the batched, padded training set results in a marginal usage of around 100 MB of RAM) would end up using 16GB or RAM. I would also like to note that calling `gc.collect` after training _does not_ empty the used RAM, which is only freed (instantly) when ending the python process.
**Describe the expected behavior**
The fitting process should go one fine, and the memory should not get saturated (I would expect some tensors to be de-allocated as iterations pass).
**Code to reproduce the issue**
The script I wrote to illustrate the issue is publicly accessible as a gist [here](https://gist.github.com/pandrey-fr/c4ba8022c5dd956388e984f49c89ce61).
**Other info / logs**
The full error stack (_with GPU enabled_) is publicly accessible as a gist [here](https://gist.github.com/pandrey-fr/ff004b4cdd6d22b9cd84f82ef4e3a5ac) | non_defect | tf keras layers embedding causes memory leak system information have i written custom code yes os platform and distribution linux mint tensorflow installed from binary using pip tensorflow version python version cuda cudnn version gpu model and memory nvidia quadro gb describe the current behavior a gpu edit cpu as well see addendum below memory leak rapidly emerges from using high dimensional tf keras layers embedding layers to be more precise i am working on transformer networks and found out that when i try to fit one e g on the portuguese to english translation task presented in a gpu memory leak emerges after a few iterations based on i rapidly came to suspect that the issue comes from the learnable embedding layers at the base of both the encoder and decoder parts of the network to further assess the issue and its source i implemented a pseudo transformer network see code linked below that is stripped of most technical components the actual model embarks e g i removed positional encoding residual connections masking mechanisms etc the rationale being to provide a more condense and faster run code to document this issue but also to confirm that the leak does not come from custom layers or any complex data processing mechanism the provided code includes a data pre processing pipeline entirely based on the aforementioned a model construction function that makes use of the keras functional api and a main function to call the former and start the fitting process on my computer everything runs fine and i can see the first few fitting iterations pass until an ugly stack of allocation error messages show up see full log linked below whose informative part seems to be w tensorflow core framework op kernel cc op requires failed at cwise ops common cc resource exhausted oom when allocating tensor addendum i re ran the provided code disabling access to the gpu and it turns out there also is a high memory usage when running on cpu during the first epoch and mostly during its first half memory usage goes up multiple gb in my case from to gb with an increase from to within the first train steps out of and keeps slowly increasing throughout the following epochs with minor decreases between increases thus displaying local plateaux which i would guess are related to the loading discarding of data batches although it is a bit less of a problem than with gpu since it is relatively common to have quite some ram available plus some swap space on linux it still does not feel right that fitting the fake model on a dataset which can be fully loaded in memory creating a list of eager tensors from the tf data dataset object containing the batched padded training set results in a marginal usage of around mb of ram would end up using or ram i would also like to note that calling gc collect after training does not empty the used ram which is only freed instantly when ending the python process describe the expected behavior the fitting process should go one fine and the memory should not get saturated i would expect some tensors to be de allocated as iterations pass code to reproduce the issue the script i wrote to illustrate the issue is publicly accessible as a gist other info logs the full error stack with gpu enabled is publicly accessible as a gist | 0 |
53,405 | 13,261,535,256 | IssuesEvent | 2020-08-20 20:04:31 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | Qt detection broken for /cvmfs/icecube.opensciencegrid.org/py2-v1 (Trac #1330) | Migrated from Trac cmake defect | This clause in config.cmake:
```text
#
# Find CVMFS software
# (designed for py2-v2 port-less software)
#
if($ENV{SROOT} MATCHES "^/cvmfs/icecube" AND NOT SYSTEM_PACKAGES)
set(CMAKE_PREFIX_PATH $ENV{SROOT})
set(SYSTEM_PACKAGES TRUE)
set(USE_CVMFS TRUE CACHE BOOL "Are we using CVMFS?")
endif()
```
unconditionally sets `SYSTEM_PACKAGES=TRUE` when using a cvmfs toolset. If the system is stock RHEL6 with Qt 4.6 installed, this will cause CMake to find the `/usr/bin/qt4-qmake` associated with 4.6 rather than the 4.8 shipped in `I3_PORTS`.
Do we really want to fall back to system search paths in this case? If so, then qt4.cmake needs to be taught how to select a minimum version of Qt rather than failing if the first one it finds is too old.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1330">https://code.icecube.wisc.edu/projects/icecube/ticket/1330</a>, reported by jvansantenand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:01:05",
"_ts": "1547251265192575",
"description": "This clause in config.cmake:\n{{{\n#\n# Find CVMFS software\n# (designed for py2-v2 port-less software)\n#\nif($ENV{SROOT} MATCHES \"^/cvmfs/icecube\" AND NOT SYSTEM_PACKAGES)\n set(CMAKE_PREFIX_PATH $ENV{SROOT})\n set(SYSTEM_PACKAGES TRUE)\n set(USE_CVMFS TRUE CACHE BOOL \"Are we using CVMFS?\")\nendif()\n}}}\n\nunconditionally sets `SYSTEM_PACKAGES=TRUE` when using a cvmfs toolset. If the system is stock RHEL6 with Qt 4.6 installed, this will cause CMake to find the `/usr/bin/qt4-qmake` associated with 4.6 rather than the 4.8 shipped in `I3_PORTS`.\n\nDo we really want to fall back to system search paths in this case? If so, then qt4.cmake needs to be taught how to select a minimum version of Qt rather than failing if the first one it finds is too old.",
"reporter": "jvansanten",
"cc": "hbretz",
"resolution": "fixed",
"time": "2015-09-04T19:46:59",
"component": "cmake",
"summary": "Qt detection broken for /cvmfs/icecube.opensciencegrid.org/py2-v1",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Qt detection broken for /cvmfs/icecube.opensciencegrid.org/py2-v1 (Trac #1330) - This clause in config.cmake:
```text
#
# Find CVMFS software
# (designed for py2-v2 port-less software)
#
if($ENV{SROOT} MATCHES "^/cvmfs/icecube" AND NOT SYSTEM_PACKAGES)
set(CMAKE_PREFIX_PATH $ENV{SROOT})
set(SYSTEM_PACKAGES TRUE)
set(USE_CVMFS TRUE CACHE BOOL "Are we using CVMFS?")
endif()
```
unconditionally sets `SYSTEM_PACKAGES=TRUE` when using a cvmfs toolset. If the system is stock RHEL6 with Qt 4.6 installed, this will cause CMake to find the `/usr/bin/qt4-qmake` associated with 4.6 rather than the 4.8 shipped in `I3_PORTS`.
Do we really want to fall back to system search paths in this case? If so, then qt4.cmake needs to be taught how to select a minimum version of Qt rather than failing if the first one it finds is too old.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1330">https://code.icecube.wisc.edu/projects/icecube/ticket/1330</a>, reported by jvansantenand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:01:05",
"_ts": "1547251265192575",
"description": "This clause in config.cmake:\n{{{\n#\n# Find CVMFS software\n# (designed for py2-v2 port-less software)\n#\nif($ENV{SROOT} MATCHES \"^/cvmfs/icecube\" AND NOT SYSTEM_PACKAGES)\n set(CMAKE_PREFIX_PATH $ENV{SROOT})\n set(SYSTEM_PACKAGES TRUE)\n set(USE_CVMFS TRUE CACHE BOOL \"Are we using CVMFS?\")\nendif()\n}}}\n\nunconditionally sets `SYSTEM_PACKAGES=TRUE` when using a cvmfs toolset. If the system is stock RHEL6 with Qt 4.6 installed, this will cause CMake to find the `/usr/bin/qt4-qmake` associated with 4.6 rather than the 4.8 shipped in `I3_PORTS`.\n\nDo we really want to fall back to system search paths in this case? If so, then qt4.cmake needs to be taught how to select a minimum version of Qt rather than failing if the first one it finds is too old.",
"reporter": "jvansanten",
"cc": "hbretz",
"resolution": "fixed",
"time": "2015-09-04T19:46:59",
"component": "cmake",
"summary": "Qt detection broken for /cvmfs/icecube.opensciencegrid.org/py2-v1",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| defect | qt detection broken for cvmfs icecube opensciencegrid org trac this clause in config cmake text find cvmfs software designed for port less software if env sroot matches cvmfs icecube and not system packages set cmake prefix path env sroot set system packages true set use cvmfs true cache bool are we using cvmfs endif unconditionally sets system packages true when using a cvmfs toolset if the system is stock with qt installed this will cause cmake to find the usr bin qmake associated with rather than the shipped in ports do we really want to fall back to system search paths in this case if so then cmake needs to be taught how to select a minimum version of qt rather than failing if the first one it finds is too old migrated from json status closed changetime ts description this clause in config cmake n n n find cvmfs software n designed for port less software n nif env sroot matches cvmfs icecube and not system packages n set cmake prefix path env sroot n set system packages true n set use cvmfs true cache bool are we using cvmfs nendif n n nunconditionally sets system packages true when using a cvmfs toolset if the system is stock with qt installed this will cause cmake to find the usr bin qmake associated with rather than the shipped in ports n ndo we really want to fall back to system search paths in this case if so then cmake needs to be taught how to select a minimum version of qt rather than failing if the first one it finds is too old reporter jvansanten cc hbretz resolution fixed time component cmake summary qt detection broken for cvmfs icecube opensciencegrid org priority normal keywords milestone owner david schultz type defect | 1 |
3,915 | 2,610,083,967 | IssuesEvent | 2015-02-26 18:25:37 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳闭合性粉刺怎么去除 | auto-migrated Priority-Medium Type-Defect | ```
深圳闭合性粉刺怎么去除【深圳韩方科颜全国热线400-869-1818��
�24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以��
�国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品�
��韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反
弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国��
�专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸�
��的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:54 | 1.0 | 深圳闭合性粉刺怎么去除 - ```
深圳闭合性粉刺怎么去除【深圳韩方科颜全国热线400-869-1818��
�24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以��
�国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品�
��韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反
弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国��
�专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸�
��的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:54 | defect | 深圳闭合性粉刺怎么去除 深圳闭合性粉刺怎么去除【 �� � 】深圳韩方科颜专业祛痘连锁机构,机构以�� �国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品� ��韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反 弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国�� �专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸� ��的痘痘。 original issue reported on code google com by szft com on may at | 1 |
419,471 | 12,223,982,753 | IssuesEvent | 2020-05-02 20:10:24 | hochreutenerl/camp-dictionary | https://api.github.com/repos/hochreutenerl/camp-dictionary | opened | Tabellarische Darstellung | enhancement feedback low priority | Ich finde es gut wenn die Begriffe immer gleich breit und schön "tabellarisch" ausgerichtet ist. Dies ist auch der Fall, ausser für die jeweils letzte Zeile, die über die Breite verteilt und nicht links ausgerichtet ist: | 1.0 | Tabellarische Darstellung - Ich finde es gut wenn die Begriffe immer gleich breit und schön "tabellarisch" ausgerichtet ist. Dies ist auch der Fall, ausser für die jeweils letzte Zeile, die über die Breite verteilt und nicht links ausgerichtet ist: | non_defect | tabellarische darstellung ich finde es gut wenn die begriffe immer gleich breit und schön tabellarisch ausgerichtet ist dies ist auch der fall ausser für die jeweils letzte zeile die über die breite verteilt und nicht links ausgerichtet ist | 0 |
8,131 | 2,611,453,667 | IssuesEvent | 2015-02-27 05:00:59 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Simplified Chinese translation Update | auto-migrated Priority-Medium Type-Defect | ```
For trunk -0.9.15
Cheers
Jiero
```
Original issue reported on code.google.com by `lilil...@gmail.com` on 8 Dec 2010 at 10:42
Attachments:
* [zh_CN.tar.gz](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-118/comment-0/zh_CN.tar.gz)
| 1.0 | Simplified Chinese translation Update - ```
For trunk -0.9.15
Cheers
Jiero
```
Original issue reported on code.google.com by `lilil...@gmail.com` on 8 Dec 2010 at 10:42
Attachments:
* [zh_CN.tar.gz](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-118/comment-0/zh_CN.tar.gz)
| defect | simplified chinese translation update for trunk cheers jiero original issue reported on code google com by lilil gmail com on dec at attachments | 1 |
139,235 | 20,815,590,120 | IssuesEvent | 2022-03-18 09:54:41 | GoogleForCreators/web-stories-wp | https://api.github.com/repos/GoogleForCreators/web-stories-wp | opened | Style tab redesign: Final style / UX polish | Type: Enhancement P1 UX Needed Group: Design Panel Pod: Prometheus Group: Workspace Type: UX Quality | <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
See [Figma section](https://www.figma.com/file/dc9xrVlRhQq4KTE9WQzU1R/_Editor---Sprint-Feature-Level-Designs_?node-id=11505%3A454596)
- [ ] Update the style of the new tabs based on Figma
- [ ] Clear out inconsistencies in Figma (e.g. some panels now don't seem to be collapsible inside the Selection tab, others still do, as of 18th of March)
- [ ] If needed, coordinate with UX to use colors and buttons etc. for the design system (if Figma doesn't use those)
- [ ] Review accessibility / keyboard navigation based on UX changes (e.g. if panels are not collapsible, we need to remove some logic + local storage logic for storing the collapsed state)
- [ ] Ensure highlighting still works after potential UX changes
<!-- A clear and concise description of what the problem is and what you want to happen. -->
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
| 1.0 | Style tab redesign: Final style / UX polish - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
See [Figma section](https://www.figma.com/file/dc9xrVlRhQq4KTE9WQzU1R/_Editor---Sprint-Feature-Level-Designs_?node-id=11505%3A454596)
- [ ] Update the style of the new tabs based on Figma
- [ ] Clear out inconsistencies in Figma (e.g. some panels now don't seem to be collapsible inside the Selection tab, others still do, as of 18th of March)
- [ ] If needed, coordinate with UX to use colors and buttons etc. for the design system (if Figma doesn't use those)
- [ ] Review accessibility / keyboard navigation based on UX changes (e.g. if panels are not collapsible, we need to remove some logic + local storage logic for storing the collapsed state)
- [ ] Ensure highlighting still works after potential UX changes
<!-- A clear and concise description of what the problem is and what you want to happen. -->
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
| non_defect | style tab redesign final style ux polish feature description see update the style of the new tabs based on figma clear out inconsistencies in figma e g some panels now don t seem to be collapsible inside the selection tab others still do as of of march if needed coordinate with ux to use colors and buttons etc for the design system if figma doesn t use those review accessibility keyboard navigation based on ux changes e g if panels are not collapsible we need to remove some logic local storage logic for storing the collapsed state ensure highlighting still works after potential ux changes alternatives considered additional context | 0 |
52,574 | 13,224,842,767 | IssuesEvent | 2020-08-17 19:57:46 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | 15k disks for icecode (Trac #7) | Migrated from Trac defect infrastructure | svn box has 2x 15k and 1x 10k disks. need 2x 15x additional, one so the 10k doesn't throttle the other two 15ks, one for a spare.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/7">https://code.icecube.wisc.edu/projects/icecube/ticket/7</a>, reported by troyand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-29T19:29:28",
"_ts": "1196364568000000",
"description": "svn box has 2x 15k and 1x 10k disks. need 2x 15x additional, one so the 10k doesn't throttle the other two 15ks, one for a spare.",
"reporter": "troy",
"cc": "",
"resolution": "invalid",
"time": "2007-06-03T16:26:55",
"component": "infrastructure",
"summary": "15k disks for icecode",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
| 1.0 | 15k disks for icecode (Trac #7) - svn box has 2x 15k and 1x 10k disks. need 2x 15x additional, one so the 10k doesn't throttle the other two 15ks, one for a spare.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/7">https://code.icecube.wisc.edu/projects/icecube/ticket/7</a>, reported by troyand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2007-11-29T19:29:28",
"_ts": "1196364568000000",
"description": "svn box has 2x 15k and 1x 10k disks. need 2x 15x additional, one so the 10k doesn't throttle the other two 15ks, one for a spare.",
"reporter": "troy",
"cc": "",
"resolution": "invalid",
"time": "2007-06-03T16:26:55",
"component": "infrastructure",
"summary": "15k disks for icecode",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
| defect | disks for icecode trac svn box has and disks need additional one so the doesn t throttle the other two one for a spare migrated from json status closed changetime ts description svn box has and disks need additional one so the doesn t throttle the other two one for a spare reporter troy cc resolution invalid time component infrastructure summary disks for icecode priority normal keywords milestone owner blaufuss type defect | 1 |
76,715 | 15,496,179,135 | IssuesEvent | 2021-03-11 02:12:23 | mwilliams7197/zendo | https://api.github.com/repos/mwilliams7197/zendo | closed | CVE-2015-9251 (Medium) detected in jquery-2.1.4.min.js, jquery-1.7.1.min.js - autoclosed | security vulnerability | ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: zendo/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: zendo/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: zendo/node_modules/sockjs/examples/multiplex/index.html</p>
<p>Path to vulnerable library: zendo/node_modules/sockjs/examples/multiplex/index.html,zendo/node_modules/sockjs/examples/echo/index.html,zendo/node_modules/sockjs/examples/hapi/html/index.html,zendo/node_modules/sockjs/examples/express-3.x/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/zendo/commit/cf147eff03fa219eea73721a78a106371bf2f0bc">cf147eff03fa219eea73721a78a106371bf2f0bc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.4","isTransitiveDependency":false,"dependencyTree":"jquery:2.1.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"}],"vulnerabilityIdentifier":"CVE-2015-9251","vulnerabilityDetails":"jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2015-9251 (Medium) detected in jquery-2.1.4.min.js, jquery-1.7.1.min.js - autoclosed - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: zendo/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: zendo/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: zendo/node_modules/sockjs/examples/multiplex/index.html</p>
<p>Path to vulnerable library: zendo/node_modules/sockjs/examples/multiplex/index.html,zendo/node_modules/sockjs/examples/echo/index.html,zendo/node_modules/sockjs/examples/hapi/html/index.html,zendo/node_modules/sockjs/examples/express-3.x/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mwilliams7197/zendo/commit/cf147eff03fa219eea73721a78a106371bf2f0bc">cf147eff03fa219eea73721a78a106371bf2f0bc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.4","isTransitiveDependency":false,"dependencyTree":"jquery:2.1.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"}],"vulnerabilityIdentifier":"CVE-2015-9251","vulnerabilityDetails":"jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_defect | cve medium detected in jquery min js jquery min js autoclosed cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file zendo node modules js attic test moment index html path to vulnerable library zendo node modules js attic test moment index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file zendo node modules sockjs examples multiplex index html path to vulnerable library zendo node modules sockjs examples multiplex index html zendo node modules sockjs examples echo index html zendo node modules sockjs examples hapi html index html zendo node modules sockjs examples express x index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed vulnerabilityurl | 0 |
138,199 | 5,329,468,083 | IssuesEvent | 2017-02-15 14:57:16 | go-gitea/gitea | https://api.github.com/repos/go-gitea/gitea | closed | Explore repositories at try.gitea.io | kind/bug priority/critical | - Gitea version (or commit ref): try.gitea.io
- Operating system: Windows 7
- Can you reproduce the bug at https://try.gitea.io:
- [X] Yes (provide example URL) https://try.gitea.io/explore/repos
## Description
**How to reproduce**
1. Surf to https://try.gitea.io/explore/repos (I tried chrome and IE)
**Expected**
2. Some sort of demo page of gitea with repositories
**Actual**
2. Text error message: template: explore/repo_list:5:200: executing "explore/repo_list" at <.Owner.Name>: can't evaluate field Name in type *models.User
Extra comment: "Because first impressions last"
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/42049999-explore-repositories-at-try-gitea-io?utm_campaign=plugin&utm_content=tracker%2F47456670&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F47456670&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Explore repositories at try.gitea.io - - Gitea version (or commit ref): try.gitea.io
- Operating system: Windows 7
- Can you reproduce the bug at https://try.gitea.io:
- [X] Yes (provide example URL) https://try.gitea.io/explore/repos
## Description
**How to reproduce**
1. Surf to https://try.gitea.io/explore/repos (I tried chrome and IE)
**Expected**
2. Some sort of demo page of gitea with repositories
**Actual**
2. Text error message: template: explore/repo_list:5:200: executing "explore/repo_list" at <.Owner.Name>: can't evaluate field Name in type *models.User
Extra comment: "Because first impressions last"
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/42049999-explore-repositories-at-try-gitea-io?utm_campaign=plugin&utm_content=tracker%2F47456670&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F47456670&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_defect | explore repositories at try gitea io gitea version or commit ref try gitea io operating system windows can you reproduce the bug at yes provide example url description how to reproduce surf to i tried chrome and ie expected some sort of demo page of gitea with repositories actual text error message template explore repo list executing explore repo list at can t evaluate field name in type models user extra comment because first impressions last want to back this issue we accept bounties via | 0 |
125,865 | 26,744,132,639 | IssuesEvent | 2023-01-30 14:56:01 | porsche-design-system/porsche-design-system | https://api.github.com/repos/porsche-design-system/porsche-design-system | closed | Use `initializeNextTick` Stencil compiler option | component code 👨🏻💻 team efficiency | Scope
---
Stencil offers a compiler option called `initializeNextTick` as seen here https://github.com/ionic-team/stencil/blob/80ef14361613706d38ad6799a03d3e5f1e1e6cbe/src/runtime/connected-callback.ts#L91-L99
In essence it is a built in workaround for delayed prop binding that occurs in Angular.
We had issues with this in the past with our form wrapper components where the child (e.g. `input`) didn't have bound attributes when `connectedCallback` was invoked. To solve this, we moved those calls into `componentWillLoad` to delay the action.
Same prop binding issues probably also apply on attributes/props on the host element as well.
Out of Scope
---
List topics that are not part of this task.
Notes
---
Share any further information that is relevant for processing.
Acceptance criteria
---
- [x] `initializeNextTick` compiler option is active
- [ ] all `componentWillLoad` lifecycles are checked and moved into `connectedCallback` if possible
Subtasks
---
- [x] apply `initializeNextTick` flag
- [x] adjust lifecycles for all components
- [x] adjust unit tests
Result
---
- `initializeNextTick` doesn't work for us within Angular since Angular's attribute rendering is still slower than our web components: https://github.com/porsche-design-system/porsche-design-system/actions/runs/3821582381/jobs/6500873167
- perhaps delaying with `requestAnimationFrame()` or `setTimeout(() => {}, 0)` might be workaround but would requirer patching stencil's core
| 1.0 | Use `initializeNextTick` Stencil compiler option - Scope
---
Stencil offers a compiler option called `initializeNextTick` as seen here https://github.com/ionic-team/stencil/blob/80ef14361613706d38ad6799a03d3e5f1e1e6cbe/src/runtime/connected-callback.ts#L91-L99
In essence it is a built in workaround for delayed prop binding that occurs in Angular.
We had issues with this in the past with our form wrapper components where the child (e.g. `input`) didn't have bound attributes when `connectedCallback` was invoked. To solve this, we moved those calls into `componentWillLoad` to delay the action.
Same prop binding issues probably also apply on attributes/props on the host element as well.
Out of Scope
---
List topics that are not part of this task.
Notes
---
Share any further information that is relevant for processing.
Acceptance criteria
---
- [x] `initializeNextTick` compiler option is active
- [ ] all `componentWillLoad` lifecycles are checked and moved into `connectedCallback` if possible
Subtasks
---
- [x] apply `initializeNextTick` flag
- [x] adjust lifecycles for all components
- [x] adjust unit tests
Result
---
- `initializeNextTick` doesn't work for us within Angular since Angular's attribute rendering is still slower than our web components: https://github.com/porsche-design-system/porsche-design-system/actions/runs/3821582381/jobs/6500873167
- perhaps delaying with `requestAnimationFrame()` or `setTimeout(() => {}, 0)` might be workaround but would requirer patching stencil's core
| non_defect | use initializenexttick stencil compiler option scope stencil offers a compiler option called initializenexttick as seen here in essence it is a built in workaround for delayed prop binding that occurs in angular we had issues with this in the past with our form wrapper components where the child e g input didn t have bound attributes when connectedcallback was invoked to solve this we moved those calls into componentwillload to delay the action same prop binding issues probably also apply on attributes props on the host element as well out of scope list topics that are not part of this task notes share any further information that is relevant for processing acceptance criteria initializenexttick compiler option is active all componentwillload lifecycles are checked and moved into connectedcallback if possible subtasks apply initializenexttick flag adjust lifecycles for all components adjust unit tests result initializenexttick doesn t work for us within angular since angular s attribute rendering is still slower than our web components perhaps delaying with requestanimationframe or settimeout might be workaround but would requirer patching stencil s core | 0 |
31,296 | 6,492,726,991 | IssuesEvent | 2017-08-21 14:28:07 | Openki/Openki | https://api.github.com/repos/Openki/Openki | closed | OAuth Login: Login with GitHub doesn't work | Defect | Bug found by @panosnethood: Using firefox (version `54.0`) and chrome a dialog window opens up but the GitHub login page doesn't load. | 1.0 | OAuth Login: Login with GitHub doesn't work - Bug found by @panosnethood: Using firefox (version `54.0`) and chrome a dialog window opens up but the GitHub login page doesn't load. | defect | oauth login login with github doesn t work bug found by panosnethood using firefox version and chrome a dialog window opens up but the github login page doesn t load | 1 |
84,199 | 10,354,637,451 | IssuesEvent | 2019-09-05 14:07:36 | vtex/styleguide | https://api.github.com/repos/vtex/styleguide | closed | Filter Bar documentation is broken | 🐛 Bug 💻 Developing... 📝 Documentation | **Describe the bug**
Filter bar preview in the documentation, both in its own documentation and in the table.
**To Reproduce**
Open the styleguide documentation and test for yourself.
**Expected behavior**
It shouldn't return any errors.
**Screenshots**


| 1.0 | Filter Bar documentation is broken - **Describe the bug**
Filter bar preview in the documentation, both in its own documentation and in the table.
**To Reproduce**
Open the styleguide documentation and test for yourself.
**Expected behavior**
It shouldn't return any errors.
**Screenshots**


| non_defect | filter bar documentation is broken describe the bug filter bar preview in the documentation both in its own documentation and in the table to reproduce open the styleguide documentation and test for yourself expected behavior it shouldn t return any errors screenshots | 0 |
70,463 | 23,178,890,779 | IssuesEvent | 2022-07-31 20:45:41 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | opened | Special characters are escaped in headings when they shouldn't be | T-Defect | ### Steps to reproduce
1. Have a chat open
2. Have the other party send you a message similar to:
```md
# <(^-^<)
```
The source of the message would look like this:
```json
{
"body": "<h1><(^-^<)</h1>\n",
"mimetype": "text/html"
}
```
### Outcome
#### What did you expect?
Android devices should unescape special characters first before rendering the heading.
For example, on Element Desktop and on GitHub it'll look like this:
# <(^-^<)
#### What happened instead?
The less-than (`<`) symbols aren't unescaped, so the `<` from when the contents was converted to HTML is used literally and the user actually sees this:
# \<(^-^\<)
### Your phone model
Fairphone 4
### Operating system version
Android
### Application version and app store
_No response_
### Homeserver
one.ems.host
### Will you send logs?
No
### Are you willing to provide a PR?
Yes | 1.0 | Special characters are escaped in headings when they shouldn't be - ### Steps to reproduce
1. Have a chat open
2. Have the other party send you a message similar to:
```md
# <(^-^<)
```
The source of the message would look like this:
```json
{
"body": "<h1><(^-^<)</h1>\n",
"mimetype": "text/html"
}
```
### Outcome
#### What did you expect?
Android devices should unescape special characters first before rendering the heading.
For example, on Element Desktop and on GitHub it'll look like this:
# <(^-^<)
#### What happened instead?
The less-than (`<`) symbols aren't unescaped, so the `<` from when the contents was converted to HTML is used literally and the user actually sees this:
# \<(^-^\<)
### Your phone model
Fairphone 4
### Operating system version
Android
### Application version and app store
_No response_
### Homeserver
one.ems.host
### Will you send logs?
No
### Are you willing to provide a PR?
Yes | defect | special characters are escaped in headings when they shouldn t be steps to reproduce have a chat open have the other party send you a message similar to md the source of the message would look like this json body lt lt n mimetype text html outcome what did you expect android devices should unescape special characters first before rendering the heading for example on element desktop and on github it ll look like this what happened instead the less than symbols aren t unescaped so the lt from when the contents was converted to html is used literally and the user actually sees this lt lt your phone model fairphone operating system version android application version and app store no response homeserver one ems host will you send logs no are you willing to provide a pr yes | 1 |
3,660 | 2,610,066,534 | IssuesEvent | 2015-02-26 18:19:33 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 临海检查前列腺炎哪里好 | auto-migrated Priority-Medium Type-Defect | ```
临海检查前列腺炎哪里好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:36 | 1.0 | 临海检查前列腺炎哪里好 - ```
临海检查前列腺炎哪里好【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:36 | defect | 临海检查前列腺炎哪里好 临海检查前列腺炎哪里好【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
41,896 | 10,697,756,352 | IssuesEvent | 2019-10-23 17:12:28 | ubc/iPeer | https://api.github.com/repos/ubc/iPeer | closed | Instructors not able to copy evaluation templates | T: defect | Instructors not able to open the context menu to copy evaluation templates. Observed Javascript errors on browser. | 1.0 | Instructors not able to copy evaluation templates - Instructors not able to open the context menu to copy evaluation templates. Observed Javascript errors on browser. | defect | instructors not able to copy evaluation templates instructors not able to open the context menu to copy evaluation templates observed javascript errors on browser | 1 |
222,268 | 17,403,430,369 | IssuesEvent | 2021-08-03 00:04:48 | Azure/azure-sdk-for-js | https://api.github.com/repos/Azure/azure-sdk-for-js | closed | [Metrics Advisor] test failing in nightly runs | Client Cognitive - Metrics Advisor test-reliability | Metrics Advisor nightly test runs are failing with:
> Error message:
> Invalid parameter. config count should be less than or equal to 10. TraceId: 22d8d9d6-e90b-4a77-a3ad-93304a59f0fb
>
> Stack trace:
> RestError: Invalid parameter. config count should be less than or equal to 10. TraceId: 22d8d9d6-e90b-4a77-a3ad-93304a59f0fb
at handleErrorResponse (/mnt/vss/_work/1/s/sdk/core/core-http/src/policies/deserializationPolicy.ts:266:17)
at /mnt/vss/_work/1/s/sdk/core/core-http/src/policies/deserializationPolicy.ts:171:47
at processTicksAndRejections (internal/process/task_queues.js:97:5)
at TracingPolicy.sendRequest (/mnt/vss/_work/1/s/sdk/core/core-http/src/policies/tracingPolicy.ts:91:24)
at GeneratedClient.sendOperationRequest (/mnt/vss/_work/1/s/sdk/core/core-http/src/serviceClient.ts:518:23)
at MetricsAdvisorAdministrationClient.createDetectionConfig (/mnt/vss/_work/1/s/sdk/metricsadvisor/ai-metrics-advisor/src/metricsAdvisorAdministrationClient.ts:560:22)
at Context.<anonymous> (/mnt/vss/_work/1/s/sdk/metricsadvisor/ai-metrics-advisor/test/public/adminclient.spec.ts:139:26)
For more details check here:
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1021375&view=results
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1019200&view=results | 1.0 | [Metrics Advisor] test failing in nightly runs - Metrics Advisor nightly test runs are failing with:
> Error message:
> Invalid parameter. config count should be less than or equal to 10. TraceId: 22d8d9d6-e90b-4a77-a3ad-93304a59f0fb
>
> Stack trace:
> RestError: Invalid parameter. config count should be less than or equal to 10. TraceId: 22d8d9d6-e90b-4a77-a3ad-93304a59f0fb
at handleErrorResponse (/mnt/vss/_work/1/s/sdk/core/core-http/src/policies/deserializationPolicy.ts:266:17)
at /mnt/vss/_work/1/s/sdk/core/core-http/src/policies/deserializationPolicy.ts:171:47
at processTicksAndRejections (internal/process/task_queues.js:97:5)
at TracingPolicy.sendRequest (/mnt/vss/_work/1/s/sdk/core/core-http/src/policies/tracingPolicy.ts:91:24)
at GeneratedClient.sendOperationRequest (/mnt/vss/_work/1/s/sdk/core/core-http/src/serviceClient.ts:518:23)
at MetricsAdvisorAdministrationClient.createDetectionConfig (/mnt/vss/_work/1/s/sdk/metricsadvisor/ai-metrics-advisor/src/metricsAdvisorAdministrationClient.ts:560:22)
at Context.<anonymous> (/mnt/vss/_work/1/s/sdk/metricsadvisor/ai-metrics-advisor/test/public/adminclient.spec.ts:139:26)
For more details check here:
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1021375&view=results
- https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1019200&view=results | non_defect | test failing in nightly runs metrics advisor nightly test runs are failing with error message invalid parameter config count should be less than or equal to traceid stack trace resterror invalid parameter config count should be less than or equal to traceid at handleerrorresponse mnt vss work s sdk core core http src policies deserializationpolicy ts at mnt vss work s sdk core core http src policies deserializationpolicy ts at processticksandrejections internal process task queues js at tracingpolicy sendrequest mnt vss work s sdk core core http src policies tracingpolicy ts at generatedclient sendoperationrequest mnt vss work s sdk core core http src serviceclient ts at metricsadvisoradministrationclient createdetectionconfig mnt vss work s sdk metricsadvisor ai metrics advisor src metricsadvisoradministrationclient ts at context mnt vss work s sdk metricsadvisor ai metrics advisor test public adminclient spec ts for more details check here | 0 |
10,063 | 2,618,934,112 | IssuesEvent | 2015-03-03 00:01:23 | chrsmith/open-ig | https://api.github.com/repos/chrsmith/open-ig | closed | game won't launch | auto-migrated Install Priority-Medium Type-Defect | ```
Game version: 0.95.130
Operating System: Windows 8 64bit
Java runtime version: 1.7.0_05
Installed using the Launcher? Yes
Game language (en, hu, de): EN
What steps will reproduce the problem?
1. Install the game using the launcher
2. Press "Run Game"
3. First time launch comes up
4. After ok nothing happens
or
1. tried launching game using open-ig-0.95.130.jar
2. "Unable to locate some game resources" using the launcher to fix it will
verify the files, but still can't launch. When no is pressed nothing happens
and this error report is on screen :
An unexpected error occurred.
You should consider submitting an error report via the project issue list:
https://code.google.com/p/open-ig/issues/list
Please include the following diagnostic information followed by the error
stacktrace(s):
Java version: 1.7.0_05
Java vendor: Oracle Corporation (http://java.oracle.com/)
Java class version: 51.0
Operating system: Windows NT (unknown), x86, 6.2
Game version: 0.95.130
Command line: [-memonce]
Available memory: 57 MB
Maximum memory: 742 MB
Parallelism: 8
Language: hu
Date and time: 2013-07-01T22:49:48.090Z
----
A crash save may have been created. Please attach it in the issue report
(zipped).
java.lang.AssertionError: Missing resource: hu ui/add_production
at hu.openig.sound.Sounds.<init>(Sounds.java:103)
at hu.openig.screen.CommonResources.init(CommonResources.java:348)
at hu.openig.screen.CommonResources.<init>(CommonResources.java:227)
at hu.openig.Startup$6.run(Startup.java:250)
at java.awt.event.InvocationEvent.dispatch(Unknown Source)
at java.awt.EventQueue.dispatchEventImpl(Unknown Source)
at java.awt.EventQueue.access$000(Unknown Source)
at java.awt.EventQueue$3.run(Unknown Source)
at java.awt.EventQueue$3.run(Unknown Source)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(Unknown Source)
at java.awt.EventQueue.dispatchEvent(Unknown Source)
at java.awt.EventDispatchThread.pumpOneEventForFilters(Unknown Source)
at java.awt.EventDispatchThread.pumpEventsForFilter(Unknown Source)
at java.awt.EventDispatchThread.pumpEventsForHierarchy(Unknown Source)
at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
at java.awt.EventDispatchThread.run(Unknown Source)
```
Original issue reported on code.google.com by `chriseil...@gmail.com` on 1 Jul 2013 at 10:53
Attachments:
* [open-ig.log](https://storage.googleapis.com/google-code-attachments/open-ig/issue-748/comment-0/open-ig.log)
| 1.0 | game won't launch - ```
Game version: 0.95.130
Operating System: Windows 8 64bit
Java runtime version: 1.7.0_05
Installed using the Launcher? Yes
Game language (en, hu, de): EN
What steps will reproduce the problem?
1. Install the game using the launcher
2. Press "Run Game"
3. First time launch comes up
4. After ok nothing happens
or
1. tried launching game using open-ig-0.95.130.jar
2. "Unable to locate some game resources" using the launcher to fix it will
verify the files, but still can't launch. When no is pressed nothing happens
and this error report is on screen :
An unexpected error occurred.
You should consider submitting an error report via the project issue list:
https://code.google.com/p/open-ig/issues/list
Please include the following diagnostic information followed by the error
stacktrace(s):
Java version: 1.7.0_05
Java vendor: Oracle Corporation (http://java.oracle.com/)
Java class version: 51.0
Operating system: Windows NT (unknown), x86, 6.2
Game version: 0.95.130
Command line: [-memonce]
Available memory: 57 MB
Maximum memory: 742 MB
Parallelism: 8
Language: hu
Date and time: 2013-07-01T22:49:48.090Z
----
A crash save may have been created. Please attach it in the issue report
(zipped).
java.lang.AssertionError: Missing resource: hu ui/add_production
at hu.openig.sound.Sounds.<init>(Sounds.java:103)
at hu.openig.screen.CommonResources.init(CommonResources.java:348)
at hu.openig.screen.CommonResources.<init>(CommonResources.java:227)
at hu.openig.Startup$6.run(Startup.java:250)
at java.awt.event.InvocationEvent.dispatch(Unknown Source)
at java.awt.EventQueue.dispatchEventImpl(Unknown Source)
at java.awt.EventQueue.access$000(Unknown Source)
at java.awt.EventQueue$3.run(Unknown Source)
at java.awt.EventQueue$3.run(Unknown Source)
at java.security.AccessController.doPrivileged(Native Method)
at java.security.ProtectionDomain$1.doIntersectionPrivilege(Unknown Source)
at java.awt.EventQueue.dispatchEvent(Unknown Source)
at java.awt.EventDispatchThread.pumpOneEventForFilters(Unknown Source)
at java.awt.EventDispatchThread.pumpEventsForFilter(Unknown Source)
at java.awt.EventDispatchThread.pumpEventsForHierarchy(Unknown Source)
at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
at java.awt.EventDispatchThread.run(Unknown Source)
```
Original issue reported on code.google.com by `chriseil...@gmail.com` on 1 Jul 2013 at 10:53
Attachments:
* [open-ig.log](https://storage.googleapis.com/google-code-attachments/open-ig/issue-748/comment-0/open-ig.log)
| defect | game won t launch game version operating system windows java runtime version installed using the launcher yes game language en hu de en what steps will reproduce the problem install the game using the launcher press run game first time launch comes up after ok nothing happens or tried launching game using open ig jar unable to locate some game resources using the launcher to fix it will verify the files but still can t launch when no is pressed nothing happens and this error report is on screen an unexpected error occurred you should consider submitting an error report via the project issue list please include the following diagnostic information followed by the error stacktrace s java version java vendor oracle corporation java class version operating system windows nt unknown game version command line available memory mb maximum memory mb parallelism language hu date and time a crash save may have been created please attach it in the issue report zipped java lang assertionerror missing resource hu ui add production at hu openig sound sounds sounds java at hu openig screen commonresources init commonresources java at hu openig screen commonresources commonresources java at hu openig startup run startup java at java awt event invocationevent dispatch unknown source at java awt eventqueue dispatcheventimpl unknown source at java awt eventqueue access unknown source at java awt eventqueue run unknown source at java awt eventqueue run unknown source at java security accesscontroller doprivileged native method at java security protectiondomain dointersectionprivilege unknown source at java awt eventqueue dispatchevent unknown source at java awt eventdispatchthread pumponeeventforfilters unknown source at java awt eventdispatchthread pumpeventsforfilter unknown source at java awt eventdispatchthread pumpeventsforhierarchy unknown source at java awt eventdispatchthread pumpevents unknown source at java awt eventdispatchthread pumpevents unknown source at java awt eventdispatchthread run unknown source original issue reported on code google com by chriseil gmail com on jul at attachments | 1 |
11,197 | 2,641,737,326 | IssuesEvent | 2015-03-11 19:27:11 | chrsmith/html5rocks | https://api.github.com/repos/chrsmith/html5rocks | closed | slides: CSS Flexible Box Layout Module | Priority-Medium Slides Type-Defect | Original [issue 100](https://code.google.com/p/html5rocks/issues/detail?id=100) created by chrsmith on 2010-07-29T00:42:52.000Z:
Reported by neonux, May 18, 2010
Adding a page about CSS3 flexbox would make the presentation even more
feature-complete than it already is :-)
http://www.w3.org/TR/css3-flexbox/
Comment 1 by ernestd@google.com, Jun 16, 2010
(No comment was entered for this change.)
Status: Accepted
Comment 2 by paulirish@google.com, Jun 24, 2010
We can reuse this slide:
http://html5apps-io2010.appspot.com/#slide36
Might be worth taking a few more things from there.. | 1.0 | slides: CSS Flexible Box Layout Module - Original [issue 100](https://code.google.com/p/html5rocks/issues/detail?id=100) created by chrsmith on 2010-07-29T00:42:52.000Z:
Reported by neonux, May 18, 2010
Adding a page about CSS3 flexbox would make the presentation even more
feature-complete than it already is :-)
http://www.w3.org/TR/css3-flexbox/
Comment 1 by ernestd@google.com, Jun 16, 2010
(No comment was entered for this change.)
Status: Accepted
Comment 2 by paulirish@google.com, Jun 24, 2010
We can reuse this slide:
http://html5apps-io2010.appspot.com/#slide36
Might be worth taking a few more things from there.. | defect | slides css flexible box layout module original created by chrsmith on reported by neonux may adding a page about flexbox would make the presentation even more feature complete than it already is comment by ernestd google com jun no comment was entered for this change status accepted comment by paulirish google com jun we can reuse this slide might be worth taking a few more things from there | 1 |
44,458 | 12,180,608,099 | IssuesEvent | 2020-04-28 12:45:34 | GraseHotspot/grase-www-portal | https://api.github.com/repos/GraseHotspot/grase-www-portal | closed | Translation for Freeradius Reply Messages (Trac #31) | Migrated from Trac Other Backend defect | Find a way to get the Radius Reply messages translated. Maybe a translation service that can be accessed via ajax? (Slow?)
Migrated from http://trac.grasehotspot.org/ticket/31
``` json
{
"status": "accepted",
"changetime": "2012-03-19T20:50:13",
"description": "Find a way to get the Radius Reply messages translated. Maybe a translation service that can be accessed via ajax? (Slow?)",
"reporter": "tim",
"cc": "",
"resolution": "",
"_ts": "1332190213597505",
"component": "Other Backend",
"summary": "Translation for Freeradius Reply Messages",
"priority": "minor",
"keywords": "",
"version": "3.4",
"time": "2011-10-10T06:52:51",
"milestone": "4.0",
"owner": "tim",
"type": "defect"
}
```
| 1.0 | Translation for Freeradius Reply Messages (Trac #31) - Find a way to get the Radius Reply messages translated. Maybe a translation service that can be accessed via ajax? (Slow?)
Migrated from http://trac.grasehotspot.org/ticket/31
``` json
{
"status": "accepted",
"changetime": "2012-03-19T20:50:13",
"description": "Find a way to get the Radius Reply messages translated. Maybe a translation service that can be accessed via ajax? (Slow?)",
"reporter": "tim",
"cc": "",
"resolution": "",
"_ts": "1332190213597505",
"component": "Other Backend",
"summary": "Translation for Freeradius Reply Messages",
"priority": "minor",
"keywords": "",
"version": "3.4",
"time": "2011-10-10T06:52:51",
"milestone": "4.0",
"owner": "tim",
"type": "defect"
}
```
| defect | translation for freeradius reply messages trac find a way to get the radius reply messages translated maybe a translation service that can be accessed via ajax slow migrated from json status accepted changetime description find a way to get the radius reply messages translated maybe a translation service that can be accessed via ajax slow reporter tim cc resolution ts component other backend summary translation for freeradius reply messages priority minor keywords version time milestone owner tim type defect | 1 |
348,196 | 31,471,322,463 | IssuesEvent | 2023-08-30 07:45:15 | Granulate/gprofiler | https://api.github.com/repos/Granulate/gprofiler | opened | tests: test_not_root fails on aarch64 | tests aarch64 | The test:
```
FAILED tests/test_preconditions.py::test_not_root - assert 2 == 1
```
The cause (inside the container):
```
b'gprofiler: Failed to open /proc/self/exe: Permission denied\n'
``` | 1.0 | tests: test_not_root fails on aarch64 - The test:
```
FAILED tests/test_preconditions.py::test_not_root - assert 2 == 1
```
The cause (inside the container):
```
b'gprofiler: Failed to open /proc/self/exe: Permission denied\n'
``` | non_defect | tests test not root fails on the test failed tests test preconditions py test not root assert the cause inside the container b gprofiler failed to open proc self exe permission denied n | 0 |
323,323 | 27,715,793,298 | IssuesEvent | 2023-03-14 16:51:12 | pandas-dev/pandas | https://api.github.com/repos/pandas-dev/pandas | closed | Series.replace fails to replace value | good first issue Needs Tests replace NA - MaskedArrays | #### Code Sample, a copy-pastable example if possible
```python
In [93]: ser.eq('nil').sum()
Out[93]: 1
In [94]: ser.replace('nil', pd.NA).eq('nil').sum()
/home/torstein/anaconda3/lib/python3.7/site-packages/pandas/core/missing.py:47: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison
mask = arr == x
Out[94]: 1
In [95]: ser.loc[(ser == 'nil').fillna(False)] = pd.NA
In [104]: ser.eq('nil').sum()
Out[104]: 0
```
#### Output of ``pd.show_versions()``
<details>
[paste the output of ``pd.show_versions()`` here below this line]
INSTALLED VERSIONS
------------------
commit : None
python : 3.7.5.final.0
python-bits : 64
OS : Linux
OS-release : 5.4.18-100.fc30.x86_64
machine : x86_64
processor : x86_64
byteorder : little
LC_ALL : None
LANG : nb_NO.UTF-8
LOCALE : nb_NO.UTF-8
pandas : 1.0.1
numpy : 1.17.3
pytz : 2019.3
dateutil : 2.8.0
pip : 19.3.1
setuptools : 41.6.0.post20191030
Cython : 0.29.13
pytest : 5.2.2
hypothesis : None
sphinx : 2.2.1
blosc : None
feather : None
xlsxwriter : 1.2.2
lxml.etree : 4.4.1
html5lib : 1.0.1
pymysql : None
psycopg2 : 2.8.4 (dt dec pq3 ext lo64)
jinja2 : 2.10.3
IPython : 7.9.0
pandas_datareader: None
bs4 : 4.8.1
bottleneck : 1.2.1
fastparquet : None
gcsfs : None
lxml.etree : 4.4.1
matplotlib : 2.2.3
numexpr : 2.7.0
odfpy : None
openpyxl : 3.0.0
pandas_gbq : None
pyarrow : 0.15.1
pytables : None
pytest : 5.2.2
pyxlsb : None
s3fs : None
scipy : 1.3.1
sqlalchemy : 1.3.10
tables : 3.5.2
tabulate : 0.8.5
xarray : None
xlrd : 1.2.0
xlwt : 1.3.0
xlsxwriter : 1.2.2
numba : 0.46.0
</details>
| 1.0 | Series.replace fails to replace value - #### Code Sample, a copy-pastable example if possible
```python
In [93]: ser.eq('nil').sum()
Out[93]: 1
In [94]: ser.replace('nil', pd.NA).eq('nil').sum()
/home/torstein/anaconda3/lib/python3.7/site-packages/pandas/core/missing.py:47: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison
mask = arr == x
Out[94]: 1
In [95]: ser.loc[(ser == 'nil').fillna(False)] = pd.NA
In [104]: ser.eq('nil').sum()
Out[104]: 0
```
#### Output of ``pd.show_versions()``
<details>
[paste the output of ``pd.show_versions()`` here below this line]
INSTALLED VERSIONS
------------------
commit : None
python : 3.7.5.final.0
python-bits : 64
OS : Linux
OS-release : 5.4.18-100.fc30.x86_64
machine : x86_64
processor : x86_64
byteorder : little
LC_ALL : None
LANG : nb_NO.UTF-8
LOCALE : nb_NO.UTF-8
pandas : 1.0.1
numpy : 1.17.3
pytz : 2019.3
dateutil : 2.8.0
pip : 19.3.1
setuptools : 41.6.0.post20191030
Cython : 0.29.13
pytest : 5.2.2
hypothesis : None
sphinx : 2.2.1
blosc : None
feather : None
xlsxwriter : 1.2.2
lxml.etree : 4.4.1
html5lib : 1.0.1
pymysql : None
psycopg2 : 2.8.4 (dt dec pq3 ext lo64)
jinja2 : 2.10.3
IPython : 7.9.0
pandas_datareader: None
bs4 : 4.8.1
bottleneck : 1.2.1
fastparquet : None
gcsfs : None
lxml.etree : 4.4.1
matplotlib : 2.2.3
numexpr : 2.7.0
odfpy : None
openpyxl : 3.0.0
pandas_gbq : None
pyarrow : 0.15.1
pytables : None
pytest : 5.2.2
pyxlsb : None
s3fs : None
scipy : 1.3.1
sqlalchemy : 1.3.10
tables : 3.5.2
tabulate : 0.8.5
xarray : None
xlrd : 1.2.0
xlwt : 1.3.0
xlsxwriter : 1.2.2
numba : 0.46.0
</details>
| non_defect | series replace fails to replace value code sample a copy pastable example if possible python in ser eq nil sum out in ser replace nil pd na eq nil sum home torstein lib site packages pandas core missing py futurewarning elementwise comparison failed returning scalar instead but in the future will perform elementwise comparison mask arr x out in ser loc pd na in ser eq nil sum out output of pd show versions installed versions commit none python final python bits os linux os release machine processor byteorder little lc all none lang nb no utf locale nb no utf pandas numpy pytz dateutil pip setuptools cython pytest hypothesis none sphinx blosc none feather none xlsxwriter lxml etree pymysql none dt dec ext ipython pandas datareader none bottleneck fastparquet none gcsfs none lxml etree matplotlib numexpr odfpy none openpyxl pandas gbq none pyarrow pytables none pytest pyxlsb none none scipy sqlalchemy tables tabulate xarray none xlrd xlwt xlsxwriter numba | 0 |
30,046 | 11,796,735,005 | IssuesEvent | 2020-03-18 11:20:55 | ocadotechnology/codeforlife-portal | https://api.github.com/repos/ocadotechnology/codeforlife-portal | closed | User with 2FA enabled isn't asked for token on login | Development bug security | **Describe the bug**
If a user sets up 2FA and then logs out, upon logging back in again, they will be redirected straight to the dashboard without having been asked the security token.
**To Reproduce**
Steps to reproduce the behaviour:
1. Login as a teacher.
2. Click 'Setup Two Factor Authentication'.
3. Follow the steps to set up 2FA.
4. Log out.
5. Log back in.
6. Observe that you weren't asked to enter the security token.
**Expected behaviour**
The website should ask you to input the security token stored on your phone upon logging in. | True | User with 2FA enabled isn't asked for token on login - **Describe the bug**
If a user sets up 2FA and then logs out, upon logging back in again, they will be redirected straight to the dashboard without having been asked the security token.
**To Reproduce**
Steps to reproduce the behaviour:
1. Login as a teacher.
2. Click 'Setup Two Factor Authentication'.
3. Follow the steps to set up 2FA.
4. Log out.
5. Log back in.
6. Observe that you weren't asked to enter the security token.
**Expected behaviour**
The website should ask you to input the security token stored on your phone upon logging in. | non_defect | user with enabled isn t asked for token on login describe the bug if a user sets up and then logs out upon logging back in again they will be redirected straight to the dashboard without having been asked the security token to reproduce steps to reproduce the behaviour login as a teacher click setup two factor authentication follow the steps to set up log out log back in observe that you weren t asked to enter the security token expected behaviour the website should ask you to input the security token stored on your phone upon logging in | 0 |
77,086 | 26,766,345,963 | IssuesEvent | 2023-01-31 10:55:07 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | opened | Pills in thread list are corrupt | T-Defect | ### Steps to reproduce
1. Go to a room with threads
2. Get someone to send some mentions in the most recent message in a thread
3. View the thread list
4. Observe that the rendered pills are visually completely:

### Outcome
#### What did you expect?
Mentions not to be broken
#### What happened instead?
Broken mentions
### Your phone model
iPhone 12 Pro Max
### Operating system version
iOS16.1
### Application version
1.9.17
### Homeserver
matrix.org
### Will you send logs?
No | 1.0 | Pills in thread list are corrupt - ### Steps to reproduce
1. Go to a room with threads
2. Get someone to send some mentions in the most recent message in a thread
3. View the thread list
4. Observe that the rendered pills are visually completely:

### Outcome
#### What did you expect?
Mentions not to be broken
#### What happened instead?
Broken mentions
### Your phone model
iPhone 12 Pro Max
### Operating system version
iOS16.1
### Application version
1.9.17
### Homeserver
matrix.org
### Will you send logs?
No | defect | pills in thread list are corrupt steps to reproduce go to a room with threads get someone to send some mentions in the most recent message in a thread view the thread list observe that the rendered pills are visually completely outcome what did you expect mentions not to be broken what happened instead broken mentions your phone model iphone pro max operating system version application version homeserver matrix org will you send logs no | 1 |
66,991 | 20,797,013,044 | IssuesEvent | 2022-03-17 10:16:35 | vector-im/element-ios | https://api.github.com/repos/vector-im/element-ios | closed | Thread list shows "encrypted message" even after the message got decrypted | T-Defect S-Minor O-Occasional A-Threads Z-Labs | ### Steps to reproduce
1. Go to an encrypted room whilst your phone is syncing
2. Go to thread list
3. Click on a thread that says "encrypted message"
4. View the thread messages displayed on screen decrypted
5. Hit the back arrow
### Outcome
#### What did you expect?
To go back to the thread list and see the root event, and last message sent in that thread showing the actual message content
#### What happened instead?
It displayed "encrypted message"
### Your phone model
iPhone 13
### Operating system version
_No response_
### Application version
1.8.6
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Thread list shows "encrypted message" even after the message got decrypted - ### Steps to reproduce
1. Go to an encrypted room whilst your phone is syncing
2. Go to thread list
3. Click on a thread that says "encrypted message"
4. View the thread messages displayed on screen decrypted
5. Hit the back arrow
### Outcome
#### What did you expect?
To go back to the thread list and see the root event, and last message sent in that thread showing the actual message content
#### What happened instead?
It displayed "encrypted message"
### Your phone model
iPhone 13
### Operating system version
_No response_
### Application version
1.8.6
### Homeserver
_No response_
### Will you send logs?
No | defect | thread list shows encrypted message even after the message got decrypted steps to reproduce go to an encrypted room whilst your phone is syncing go to thread list click on a thread that says encrypted message view the thread messages displayed on screen decrypted hit the back arrow outcome what did you expect to go back to the thread list and see the root event and last message sent in that thread showing the actual message content what happened instead it displayed encrypted message your phone model iphone operating system version no response application version homeserver no response will you send logs no | 1 |
317,518 | 9,666,272,960 | IssuesEvent | 2019-05-21 10:24:10 | oceanprotocol/ocean | https://api.github.com/repos/oceanprotocol/ocean | opened | Review of the Keeper Documentation before Ocean production release | Keeper priority:medium | In order to prepare the network deployment in production and security audits, it would be necessary to update some existing documentation:
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/432
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/468
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/469
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/470
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/475
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/489 | 1.0 | Review of the Keeper Documentation before Ocean production release - In order to prepare the network deployment in production and security audits, it would be necessary to update some existing documentation:
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/432
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/468
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/469
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/470
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/475
- [ ] https://github.com/oceanprotocol/keeper-contracts/issues/489 | non_defect | review of the keeper documentation before ocean production release in order to prepare the network deployment in production and security audits it would be necessary to update some existing documentation | 0 |
249,237 | 18,858,174,507 | IssuesEvent | 2021-11-12 09:28:05 | greyguy21/pe | https://api.github.com/repos/greyguy21/pe | opened | title of use case does not match mss | severity.Medium type.DocumentationBug | title of use case should adding event to a blocked slot instead.

<!--session: 1636705515500-7fac9a9e-1533-4df4-91b1-951f4f5b06b8-->
<!--Version: Web v3.4.1--> | 1.0 | title of use case does not match mss - title of use case should adding event to a blocked slot instead.

<!--session: 1636705515500-7fac9a9e-1533-4df4-91b1-951f4f5b06b8-->
<!--Version: Web v3.4.1--> | non_defect | title of use case does not match mss title of use case should adding event to a blocked slot instead | 0 |
272,074 | 20,732,344,899 | IssuesEvent | 2022-03-14 10:35:04 | marinasainz/go-bees | https://api.github.com/repos/marinasainz/go-bees | closed | Pasar anexos a latex | documentation | - [x] software plan del proyecto
- [x] Especificación de Requisitos
- [x] Especificación de diseño
- [x] manual del programador
- [x] manual de usuario
Expresión regular para corregir las referencias:
Buscar:(\{\[\})([\w:\\]+)(\{\]\}\\\_)
Reemplazar por:\\citep{\2}
Ej: {[}art:campbell2005{]}\_->\citep{art:campbell2005}
Expresión regular para corregir las imágenes:
Buscar:(\\includegraphics{..\/..\/img\/)(\w*)(.\w*\})
Reemplazar por:\\imagen{\2}{descripción...}
Ej: \includegraphics{../../img/s2.png}->\imagen{s2}{descripción...}} | 1.0 | Pasar anexos a latex - - [x] software plan del proyecto
- [x] Especificación de Requisitos
- [x] Especificación de diseño
- [x] manual del programador
- [x] manual de usuario
Expresión regular para corregir las referencias:
Buscar:(\{\[\})([\w:\\]+)(\{\]\}\\\_)
Reemplazar por:\\citep{\2}
Ej: {[}art:campbell2005{]}\_->\citep{art:campbell2005}
Expresión regular para corregir las imágenes:
Buscar:(\\includegraphics{..\/..\/img\/)(\w*)(.\w*\})
Reemplazar por:\\imagen{\2}{descripción...}
Ej: \includegraphics{../../img/s2.png}->\imagen{s2}{descripción...}} | non_defect | pasar anexos a latex software plan del proyecto especificación de requisitos especificación de diseño manual del programador manual de usuario expresión regular para corregir las referencias buscar reemplazar por citep ej citep art expresión regular para corregir las imágenes buscar includegraphics img w w reemplazar por imagen descripción ej includegraphics img png imagen descripción | 0 |
69,035 | 22,061,873,581 | IssuesEvent | 2022-05-30 19:08:35 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | Docs suggest scipy.sparse.linalg.expm_multiply supports LinearOperator but it doesn't | defect scipy.sparse.linalg | According to [the docs](http://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.expm_multiply.html#scipy.sparse.linalg.expm_multiply) the first parameter to expm_multiply is a "transposable linear operator", which presumably includes LinearOperator. However expm_multiply actually appears to accept numpy arrays or scipy sparse matrices only. For example:
``` python
import numpy as np
import scipy.sparse.linalg as spsl
spsl.expm_multiply(spsl.aslinearoperator(np.array([[2]])), np.array([1]))
```
fails because [_trace()](https://github.com/scipy/scipy/blob/v0.16.0/scipy/sparse/linalg/_expm_multiply.py#L31) assumes its parameter is either a numpy array or a scipy sparse matrix.
It seems expm_multiply uses the trace, 1-norm and infinity-norm of its parameter, none of which are available for LinearOperator. <del>Fixing this bug by supporting LinearOperator would therefore require some non-trivial algorithm design so you may want to fix the documentation now and defer adding LinearOperator support until later.</del>
| 1.0 | Docs suggest scipy.sparse.linalg.expm_multiply supports LinearOperator but it doesn't - According to [the docs](http://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.expm_multiply.html#scipy.sparse.linalg.expm_multiply) the first parameter to expm_multiply is a "transposable linear operator", which presumably includes LinearOperator. However expm_multiply actually appears to accept numpy arrays or scipy sparse matrices only. For example:
``` python
import numpy as np
import scipy.sparse.linalg as spsl
spsl.expm_multiply(spsl.aslinearoperator(np.array([[2]])), np.array([1]))
```
fails because [_trace()](https://github.com/scipy/scipy/blob/v0.16.0/scipy/sparse/linalg/_expm_multiply.py#L31) assumes its parameter is either a numpy array or a scipy sparse matrix.
It seems expm_multiply uses the trace, 1-norm and infinity-norm of its parameter, none of which are available for LinearOperator. <del>Fixing this bug by supporting LinearOperator would therefore require some non-trivial algorithm design so you may want to fix the documentation now and defer adding LinearOperator support until later.</del>
| defect | docs suggest scipy sparse linalg expm multiply supports linearoperator but it doesn t according to the first parameter to expm multiply is a transposable linear operator which presumably includes linearoperator however expm multiply actually appears to accept numpy arrays or scipy sparse matrices only for example python import numpy as np import scipy sparse linalg as spsl spsl expm multiply spsl aslinearoperator np array np array fails because assumes its parameter is either a numpy array or a scipy sparse matrix it seems expm multiply uses the trace norm and infinity norm of its parameter none of which are available for linearoperator fixing this bug by supporting linearoperator would therefore require some non trivial algorithm design so you may want to fix the documentation now and defer adding linearoperator support until later | 1 |
71,495 | 23,655,082,837 | IssuesEvent | 2022-08-26 10:23:58 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | 'Empty Room (was %user)' is not translated | T-Defect S-Minor A-Room-List A-Room-Settings A-Room-View I18n Help Wanted O-Frequent | I believe this string comes from the js-sdk, so might need to use `td()` and such to get it properly fixed | 1.0 | 'Empty Room (was %user)' is not translated - I believe this string comes from the js-sdk, so might need to use `td()` and such to get it properly fixed | defect | empty room was user is not translated i believe this string comes from the js sdk so might need to use td and such to get it properly fixed | 1 |
704,394 | 24,195,438,831 | IssuesEvent | 2022-09-23 22:45:50 | ArctosDB/arctos | https://api.github.com/repos/ArctosDB/arctos | closed | print stuff not working | Priority-Critical (Arctos is broken) Help wanted Error Messages | I have a batch of labels to print; from the search results, I went to the manage dropdown, selected "Print any Report"; Arctos opens a new page and I clicked on the link to continue to the Arctos reporter. The page tries to load for a bit, then I got the following message:
The connection was reset
The connection to the server was reset while the page was loading.
The site could be temporarily unavailable or too busy. Try again in a few moments.
If you are unable to load any pages, check your computer’s network connection.
If your computer or network is protected by a firewall or proxy, make sure that Firefox is permitted to access the Web.
I tried to run the report in Chrome and got a similar message. I tried running troubleshoot on my computer and it came back with no issues on my end (at least with my computer). | 1.0 | print stuff not working - I have a batch of labels to print; from the search results, I went to the manage dropdown, selected "Print any Report"; Arctos opens a new page and I clicked on the link to continue to the Arctos reporter. The page tries to load for a bit, then I got the following message:
The connection was reset
The connection to the server was reset while the page was loading.
The site could be temporarily unavailable or too busy. Try again in a few moments.
If you are unable to load any pages, check your computer’s network connection.
If your computer or network is protected by a firewall or proxy, make sure that Firefox is permitted to access the Web.
I tried to run the report in Chrome and got a similar message. I tried running troubleshoot on my computer and it came back with no issues on my end (at least with my computer). | non_defect | print stuff not working i have a batch of labels to print from the search results i went to the manage dropdown selected print any report arctos opens a new page and i clicked on the link to continue to the arctos reporter the page tries to load for a bit then i got the following message the connection was reset the connection to the server was reset while the page was loading the site could be temporarily unavailable or too busy try again in a few moments if you are unable to load any pages check your computer’s network connection if your computer or network is protected by a firewall or proxy make sure that firefox is permitted to access the web i tried to run the report in chrome and got a similar message i tried running troubleshoot on my computer and it came back with no issues on my end at least with my computer | 0 |
397,760 | 27,176,353,766 | IssuesEvent | 2023-02-18 03:00:54 | LibertyDSNP/frequency | https://api.github.com/repos/LibertyDSNP/frequency | closed | Schema Creation Limits | documentation enhancement project | As Frequency, we don't want just anyone able to register schemas.
## Acceptance Criteria
- Only a simple majority frequency council member can approve a Schema Creation
- Local and rococo also allow anyone to do Schema Creation
- Documentation updated to reflect
## Notes
This is the same as #321 with these exceptions:
- Needs approval of a simple majority of the Frequency Council
- Does NOT matter who is asking for the schema to be created (aka no MSA needed or proof)
- Is likely somewhat longer term in solution
- More ok for users to go through the referendum process, although that is slow for protocols that have a lot of schemas.
| 1.0 | Schema Creation Limits - As Frequency, we don't want just anyone able to register schemas.
## Acceptance Criteria
- Only a simple majority frequency council member can approve a Schema Creation
- Local and rococo also allow anyone to do Schema Creation
- Documentation updated to reflect
## Notes
This is the same as #321 with these exceptions:
- Needs approval of a simple majority of the Frequency Council
- Does NOT matter who is asking for the schema to be created (aka no MSA needed or proof)
- Is likely somewhat longer term in solution
- More ok for users to go through the referendum process, although that is slow for protocols that have a lot of schemas.
| non_defect | schema creation limits as frequency we don t want just anyone able to register schemas acceptance criteria only a simple majority frequency council member can approve a schema creation local and rococo also allow anyone to do schema creation documentation updated to reflect notes this is the same as with these exceptions needs approval of a simple majority of the frequency council does not matter who is asking for the schema to be created aka no msa needed or proof is likely somewhat longer term in solution more ok for users to go through the referendum process although that is slow for protocols that have a lot of schemas | 0 |
746,565 | 26,035,381,511 | IssuesEvent | 2022-12-22 04:02:45 | sovware/directorist | https://api.github.com/repos/sovware/directorist | closed | Tag Icon | Type: Bug Priority: High | ### Directorist version
7.4.3
### Environment
_No response_
### Description
The tag Icon can not be changed.
Screenshot, https://prnt.sc/7S5UR-5gYjcJ
### Steps to reproduce
(1) Directory Listings> Directory Builder> Single page layout> Content> Tag
(2) Try to change the icon
(3) Icon will not be changed in the frontend
### Isolating the problem
- [X] I have reproduced this bug in localhost.
- [X] This bug happens with a default WordPress theme active (eg. 2021, 2022 etc). | 1.0 | Tag Icon - ### Directorist version
7.4.3
### Environment
_No response_
### Description
The tag Icon can not be changed.
Screenshot, https://prnt.sc/7S5UR-5gYjcJ
### Steps to reproduce
(1) Directory Listings> Directory Builder> Single page layout> Content> Tag
(2) Try to change the icon
(3) Icon will not be changed in the frontend
### Isolating the problem
- [X] I have reproduced this bug in localhost.
- [X] This bug happens with a default WordPress theme active (eg. 2021, 2022 etc). | non_defect | tag icon directorist version environment no response description the tag icon can not be changed screenshot steps to reproduce directory listings directory builder single page layout content tag try to change the icon icon will not be changed in the frontend isolating the problem i have reproduced this bug in localhost this bug happens with a default wordpress theme active eg etc | 0 |
7,642 | 2,610,408,331 | IssuesEvent | 2015-02-26 20:12:42 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | opened | [Enhancement] Num Pad | auto-migrated Priority-Medium Type-Defect | ```
Hero taunts, using the number pad, use FOC audio.
Swap out for RAW audio content.
```
-----
Original issue reported on code.google.com by `Klaku...@gmail.com` on 10 Jul 2012 at 6:35 | 1.0 | [Enhancement] Num Pad - ```
Hero taunts, using the number pad, use FOC audio.
Swap out for RAW audio content.
```
-----
Original issue reported on code.google.com by `Klaku...@gmail.com` on 10 Jul 2012 at 6:35 | defect | num pad hero taunts using the number pad use foc audio swap out for raw audio content original issue reported on code google com by klaku gmail com on jul at | 1 |
64,799 | 18,908,638,081 | IssuesEvent | 2021-11-16 11:48:49 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Maximised widgets should be restored when switching back into a room | T-Defect A-Widgets Z-Maximised-Widgets | ### Steps to reproduce
1. Maximise a widget in some room (without sharing the room layout with everyone, so local only)
2. Switch rooms and come back again
### Expected
The widget should reappear as maximised when coming back the room.
### Actual
The widget was not displayed even though the info panel said it was maximised. | 1.0 | Maximised widgets should be restored when switching back into a room - ### Steps to reproduce
1. Maximise a widget in some room (without sharing the room layout with everyone, so local only)
2. Switch rooms and come back again
### Expected
The widget should reappear as maximised when coming back the room.
### Actual
The widget was not displayed even though the info panel said it was maximised. | defect | maximised widgets should be restored when switching back into a room steps to reproduce maximise a widget in some room without sharing the room layout with everyone so local only switch rooms and come back again expected the widget should reappear as maximised when coming back the room actual the widget was not displayed even though the info panel said it was maximised | 1 |
42,601 | 11,166,340,319 | IssuesEvent | 2019-12-27 12:54:34 | hazelcast/hazelcast-nodejs-client | https://api.github.com/repos/hazelcast/hazelcast-nodejs-client | closed | [hzcloud] nodejs client should ask for new ip addresses in case of connection timeout | Type: Defect | in cloud, when we make stop and start then public ip addresses of instances will be changed. ( %90 probability)
so nodejs client should ask continuously to hazelcast cloud endpoint for the new ip addresses.
current behaviour: it is asking for 1 time after initial connection timeout.
expected behaviour: it should ask every 10 seconds or any other short interval for the new possible ip addresses. | 1.0 | [hzcloud] nodejs client should ask for new ip addresses in case of connection timeout - in cloud, when we make stop and start then public ip addresses of instances will be changed. ( %90 probability)
so nodejs client should ask continuously to hazelcast cloud endpoint for the new ip addresses.
current behaviour: it is asking for 1 time after initial connection timeout.
expected behaviour: it should ask every 10 seconds or any other short interval for the new possible ip addresses. | defect | nodejs client should ask for new ip addresses in case of connection timeout in cloud when we make stop and start then public ip addresses of instances will be changed probability so nodejs client should ask continuously to hazelcast cloud endpoint for the new ip addresses current behaviour it is asking for time after initial connection timeout expected behaviour it should ask every seconds or any other short interval for the new possible ip addresses | 1 |
53,614 | 13,261,972,725 | IssuesEvent | 2020-08-20 20:52:22 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | port install root_5.34.18 under Ubuntu 16.04 fails (Trac #1753) | Migrated from Trac defect tools/ports | Root 5.34.18, a dependency of genie 2.8.6, fails to build using !DarwinPorts:
```text
checking for the pthreads library -llthread...
bin/rmkdepend -R -fcint/reflex/src/GenreflexMemberBuilder.d -Y -w 1000 -- -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -D__cplusplus -- /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx
g++ -O2 -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -o cint/reflex/src/GenreflexMemberBuilder.o -c /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx
In file included from cint/cint/src/dict/libstrm.h:37:0,
from cint/cint/src/dict/libstrm.cxx:14:
cint/cint/lib/stream/iostrm.h:24:22: fatal error: iostream.h: No such file or directory
compilation terminated.
Makefile:616: recipe for target 'cint/cint/src/dict/libstrm.o' failed
make[1]: *** [cint/cint/src/dict/libstrm.o] Error 1
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1753">https://code.icecube.wisc.edu/projects/icecube/ticket/1753</a>, reported by jlanfranchiand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-21T18:10:44",
"_ts": "1466532644311165",
"description": "Root 5.34.18, a dependency of genie 2.8.6, fails to build using !DarwinPorts:\n\n{{{\nchecking for the pthreads library -llthread... \nbin/rmkdepend -R -fcint/reflex/src/GenreflexMemberBuilder.d -Y -w 1000 -- -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -D__cplusplus -- /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx\ng++ -O2 -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -o cint/reflex/src/GenreflexMemberBuilder.o -c /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx\nIn file included from cint/cint/src/dict/libstrm.h:37:0,\n from cint/cint/src/dict/libstrm.cxx:14:\ncint/cint/lib/stream/iostrm.h:24:22: fatal error: iostream.h: No such file or directory\ncompilation terminated.\nMakefile:616: recipe for target 'cint/cint/src/dict/libstrm.o' failed\nmake[1]: *** [cint/cint/src/dict/libstrm.o] Error 1\n}}}",
"reporter": "jlanfranchi",
"cc": "",
"resolution": "fixed",
"time": "2016-06-21T00:39:23",
"component": "tools/ports",
"summary": "port install root_5.34.18 under Ubuntu 16.04 fails",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | port install root_5.34.18 under Ubuntu 16.04 fails (Trac #1753) - Root 5.34.18, a dependency of genie 2.8.6, fails to build using !DarwinPorts:
```text
checking for the pthreads library -llthread...
bin/rmkdepend -R -fcint/reflex/src/GenreflexMemberBuilder.d -Y -w 1000 -- -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -D__cplusplus -- /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx
g++ -O2 -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -o cint/reflex/src/GenreflexMemberBuilder.o -c /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx
In file included from cint/cint/src/dict/libstrm.h:37:0,
from cint/cint/src/dict/libstrm.cxx:14:
cint/cint/lib/stream/iostrm.h:24:22: fatal error: iostream.h: No such file or directory
compilation terminated.
Makefile:616: recipe for target 'cint/cint/src/dict/libstrm.o' failed
make[1]: *** [cint/cint/src/dict/libstrm.o] Error 1
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1753">https://code.icecube.wisc.edu/projects/icecube/ticket/1753</a>, reported by jlanfranchiand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-06-21T18:10:44",
"_ts": "1466532644311165",
"description": "Root 5.34.18, a dependency of genie 2.8.6, fails to build using !DarwinPorts:\n\n{{{\nchecking for the pthreads library -llthread... \nbin/rmkdepend -R -fcint/reflex/src/GenreflexMemberBuilder.d -Y -w 1000 -- -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -D__cplusplus -- /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx\ng++ -O2 -pipe -m64 -Wshadow -Wall -W -Woverloaded-virtual -fPIC -Iinclude -pthread -o cint/reflex/src/GenreflexMemberBuilder.o -c /home/icecube/bin/ports/var/db/dports/build/file._home_icecube_bin_ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_science_root_5.34.18/work/root/cint/reflex/src/GenreflexMemberBuilder.cxx\nIn file included from cint/cint/src/dict/libstrm.h:37:0,\n from cint/cint/src/dict/libstrm.cxx:14:\ncint/cint/lib/stream/iostrm.h:24:22: fatal error: iostream.h: No such file or directory\ncompilation terminated.\nMakefile:616: recipe for target 'cint/cint/src/dict/libstrm.o' failed\nmake[1]: *** [cint/cint/src/dict/libstrm.o] Error 1\n}}}",
"reporter": "jlanfranchi",
"cc": "",
"resolution": "fixed",
"time": "2016-06-21T00:39:23",
"component": "tools/ports",
"summary": "port install root_5.34.18 under Ubuntu 16.04 fails",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | port install root under ubuntu fails trac root a dependency of genie fails to build using darwinports text checking for the pthreads library llthread bin rmkdepend r fcint reflex src genreflexmemberbuilder d y w pipe wshadow wall w woverloaded virtual fpic iinclude pthread d cplusplus home icecube bin ports var db dports build file home icecube bin ports var db dports sources rsync code icecube wisc edu icecube tools ports science root work root cint reflex src genreflexmemberbuilder cxx g pipe wshadow wall w woverloaded virtual fpic iinclude pthread o cint reflex src genreflexmemberbuilder o c home icecube bin ports var db dports build file home icecube bin ports var db dports sources rsync code icecube wisc edu icecube tools ports science root work root cint reflex src genreflexmemberbuilder cxx in file included from cint cint src dict libstrm h from cint cint src dict libstrm cxx cint cint lib stream iostrm h fatal error iostream h no such file or directory compilation terminated makefile recipe for target cint cint src dict libstrm o failed make error migrated from json status closed changetime ts description root a dependency of genie fails to build using darwinports n n nchecking for the pthreads library llthread nbin rmkdepend r fcint reflex src genreflexmemberbuilder d y w pipe wshadow wall w woverloaded virtual fpic iinclude pthread d cplusplus home icecube bin ports var db dports build file home icecube bin ports var db dports sources rsync code icecube wisc edu icecube tools ports science root work root cint reflex src genreflexmemberbuilder cxx ng pipe wshadow wall w woverloaded virtual fpic iinclude pthread o cint reflex src genreflexmemberbuilder o c home icecube bin ports var db dports build file home icecube bin ports var db dports sources rsync code icecube wisc edu icecube tools ports science root work root cint reflex src genreflexmemberbuilder cxx nin file included from cint cint src dict libstrm h n from cint cint src dict libstrm cxx ncint cint lib stream iostrm h fatal error iostream h no such file or directory ncompilation terminated nmakefile recipe for target cint cint src dict libstrm o failed nmake error n reporter jlanfranchi cc resolution fixed time component tools ports summary port install root under ubuntu fails priority normal keywords milestone owner nega type defect | 1 |
29,925 | 5,959,225,691 | IssuesEvent | 2017-05-29 10:20:29 | line/armeria | https://api.github.com/repos/line/armeria | closed | Future returned by HttpRequest.aggregate() may not complete if request times out | defect | The internal subscriber implementation of `HttpRequest.aggregate()` and `HttpResponse.aggregate()` operation do not complete the returned `CompletableFuture` when the subscription has been cancelled, because a `StreamMessage` will not call `Subscriber.onComplete()` or `onError()` when the subscription is cancelled. We need to rely on `StreamMessage.closeFuture()`. | 1.0 | Future returned by HttpRequest.aggregate() may not complete if request times out - The internal subscriber implementation of `HttpRequest.aggregate()` and `HttpResponse.aggregate()` operation do not complete the returned `CompletableFuture` when the subscription has been cancelled, because a `StreamMessage` will not call `Subscriber.onComplete()` or `onError()` when the subscription is cancelled. We need to rely on `StreamMessage.closeFuture()`. | defect | future returned by httprequest aggregate may not complete if request times out the internal subscriber implementation of httprequest aggregate and httpresponse aggregate operation do not complete the returned completablefuture when the subscription has been cancelled because a streammessage will not call subscriber oncomplete or onerror when the subscription is cancelled we need to rely on streammessage closefuture | 1 |
818,233 | 30,679,756,371 | IssuesEvent | 2023-07-26 08:22:45 | Shelf-nu/shelf.nu | https://api.github.com/repos/Shelf-nu/shelf.nu | closed | Ensure consistent font sizes for input fields | bug High Priority | We have noticed that we have some inconsistent input font sizes across different inputs.
We need to check the all and confirm all inputs including custom ones like the select from Radix, have a 16px font size. | 1.0 | Ensure consistent font sizes for input fields - We have noticed that we have some inconsistent input font sizes across different inputs.
We need to check the all and confirm all inputs including custom ones like the select from Radix, have a 16px font size. | non_defect | ensure consistent font sizes for input fields we have noticed that we have some inconsistent input font sizes across different inputs we need to check the all and confirm all inputs including custom ones like the select from radix have a font size | 0 |
3,234 | 2,610,058,721 | IssuesEvent | 2015-02-26 18:17:20 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 路桥最好的男科医院 | auto-migrated Priority-Medium Type-Defect | ```
路桥最好的男科医院【台州五洲生殖医院】24小时健康咨询热
线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒
江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118�
��198及椒江一金清公交车直达枫南小区,乘坐107、105、109、112
、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 6:47 | 1.0 | 路桥最好的男科医院 - ```
路桥最好的男科医院【台州五洲生殖医院】24小时健康咨询热
线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州市椒
江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108、118�
��198及椒江一金清公交车直达枫南小区,乘坐107、105、109、112
、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 6:47 | defect | 路桥最好的男科医院 路桥最好的男科医院【台州五洲生殖医院】 线 微信号tzwzszyy 医院地址 台州市椒 (枫南大转盘旁)乘车线路 、 、 � �� , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
785,284 | 27,607,454,736 | IssuesEvent | 2023-03-09 13:55:28 | cds-snc/resources-ressources | https://api.github.com/repos/cds-snc/resources-ressources | closed | [Jump-to] Anchor links don't show header title when jumping to it on mobile or narrow desktop screens | BUG HIGH PRIORITY | # Describe the bug
When you click on an anchor link that is expected to jump you to a specific section on the page, it does not show the header.
## To Reproduce
Steps to reproduce the behavior:
1. Go to any resource page
2. Click on the second item in the list of headings
3. See error, you have to scroll up to see the heading you clicked
## Expected behavior
User expects to land on the section heading -- not mid-section.
## Screenshots
** Clicking Alpha in "On this page" component**

** Land me in the middle of the section without visibility of the H2 **

<img width="516" alt="Screenshot 2022-10-27 at 8 18 06 AM" src="https://user-images.githubusercontent.com/38330843/198282549-e431611c-164e-4d12-8af9-00b7253db92c.png">
**Smartphone:**
- Device: iPhone 12 mini
- OS: iOS
- Browser: Google
**Desktop:**
- Device: MacBook
- OS: Ventura
- Browser: Google
- Window: Narrowed to 1/3 of the screen
**Additional context**
Found this bug while validating bug #120
Wonder if it's also relevant on desktop?
| 1.0 | [Jump-to] Anchor links don't show header title when jumping to it on mobile or narrow desktop screens - # Describe the bug
When you click on an anchor link that is expected to jump you to a specific section on the page, it does not show the header.
## To Reproduce
Steps to reproduce the behavior:
1. Go to any resource page
2. Click on the second item in the list of headings
3. See error, you have to scroll up to see the heading you clicked
## Expected behavior
User expects to land on the section heading -- not mid-section.
## Screenshots
** Clicking Alpha in "On this page" component**

** Land me in the middle of the section without visibility of the H2 **

<img width="516" alt="Screenshot 2022-10-27 at 8 18 06 AM" src="https://user-images.githubusercontent.com/38330843/198282549-e431611c-164e-4d12-8af9-00b7253db92c.png">
**Smartphone:**
- Device: iPhone 12 mini
- OS: iOS
- Browser: Google
**Desktop:**
- Device: MacBook
- OS: Ventura
- Browser: Google
- Window: Narrowed to 1/3 of the screen
**Additional context**
Found this bug while validating bug #120
Wonder if it's also relevant on desktop?
| non_defect | anchor links don t show header title when jumping to it on mobile or narrow desktop screens describe the bug when you click on an anchor link that is expected to jump you to a specific section on the page it does not show the header to reproduce steps to reproduce the behavior go to any resource page click on the second item in the list of headings see error you have to scroll up to see the heading you clicked expected behavior user expects to land on the section heading not mid section screenshots clicking alpha in on this page component land me in the middle of the section without visibility of the img width alt screenshot at am src smartphone device iphone mini os ios browser google desktop device macbook os ventura browser google window narrowed to of the screen additional context found this bug while validating bug wonder if it s also relevant on desktop | 0 |
75,683 | 25,996,396,678 | IssuesEvent | 2022-12-20 11:57:53 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | opened | selectOneMenu: Wrongly escaping input when selecting a value from the list | :lady_beetle: defect :bangbang: needs-triage | ### Describe the bug
If the selectOneMenu is used with items that contains characters that have to be escaped within the HTML than those characters are escaped wrongly within the input field of the selectOneMenu. Initially the values are correct, the error occurs if the js script is triggered for setting the text of the current selected item. The error entered with PF11 where within the "PrimeFaces.widget.SelectOneMenu.prototype.renderSelectItem" method an escaping was added "PrimeFaces.escapeHTML(..)". A reproducer is attached.
[Uploading SelectOneMenuReproducer.zip…]()
### Reproducer
Start the reproducer and open the browser on "http://localhost:8080/primefaces-test/". There you see the select one menu with the correct text. As soon as you click on the drop-down-button the text is displayed with escaped characters. Within the test.js file is a TODO where you can find the line that is causing this issue and also the line that is fixing it.
[SelectOneMenuReproducer.zip](https://github.com/primefaces/primefaces/files/10268130/SelectOneMenuReproducer.zip)
### Expected behavior
Even if another item is selected about the drop down box the displayed text should not contain escaped characters.
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
Diamond
### JSF implementation
MyFaces
### JSF version
2.3
### Java version
11
### Browser(s)
_No response_ | 1.0 | selectOneMenu: Wrongly escaping input when selecting a value from the list - ### Describe the bug
If the selectOneMenu is used with items that contains characters that have to be escaped within the HTML than those characters are escaped wrongly within the input field of the selectOneMenu. Initially the values are correct, the error occurs if the js script is triggered for setting the text of the current selected item. The error entered with PF11 where within the "PrimeFaces.widget.SelectOneMenu.prototype.renderSelectItem" method an escaping was added "PrimeFaces.escapeHTML(..)". A reproducer is attached.
[Uploading SelectOneMenuReproducer.zip…]()
### Reproducer
Start the reproducer and open the browser on "http://localhost:8080/primefaces-test/". There you see the select one menu with the correct text. As soon as you click on the drop-down-button the text is displayed with escaped characters. Within the test.js file is a TODO where you can find the line that is causing this issue and also the line that is fixing it.
[SelectOneMenuReproducer.zip](https://github.com/primefaces/primefaces/files/10268130/SelectOneMenuReproducer.zip)
### Expected behavior
Even if another item is selected about the drop down box the displayed text should not contain escaped characters.
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
Diamond
### JSF implementation
MyFaces
### JSF version
2.3
### Java version
11
### Browser(s)
_No response_ | defect | selectonemenu wrongly escaping input when selecting a value from the list describe the bug if the selectonemenu is used with items that contains characters that have to be escaped within the html than those characters are escaped wrongly within the input field of the selectonemenu initially the values are correct the error occurs if the js script is triggered for setting the text of the current selected item the error entered with where within the primefaces widget selectonemenu prototype renderselectitem method an escaping was added primefaces escapehtml a reproducer is attached reproducer start the reproducer and open the browser on there you see the select one menu with the correct text as soon as you click on the drop down button the text is displayed with escaped characters within the test js file is a todo where you can find the line that is causing this issue and also the line that is fixing it expected behavior even if another item is selected about the drop down box the displayed text should not contain escaped characters primefaces edition community primefaces version theme diamond jsf implementation myfaces jsf version java version browser s no response | 1 |
17,153 | 11,733,244,458 | IssuesEvent | 2020-03-11 06:32:26 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Auto-completion in comments not entirely useful for apostrophes. | bug topic:editor usability | **Godot version:** 3.2.0 stable
**OS/device including version:** N/A
**Issue description:** Auto-completion has its issues when you're writing comments. For example `# Can't and won't do this.` would duplicate the `'` character (i.e. treat it as a single quote). This is a minor issue, but makes contractions (i.e. "it's", "can't" etc) slightly more difficult than it needs to be. Maybe look at the way other projects such as Inform 7 (for example) handle this?
**Steps to reproduce:** Just press `'`. | True | Auto-completion in comments not entirely useful for apostrophes. - **Godot version:** 3.2.0 stable
**OS/device including version:** N/A
**Issue description:** Auto-completion has its issues when you're writing comments. For example `# Can't and won't do this.` would duplicate the `'` character (i.e. treat it as a single quote). This is a minor issue, but makes contractions (i.e. "it's", "can't" etc) slightly more difficult than it needs to be. Maybe look at the way other projects such as Inform 7 (for example) handle this?
**Steps to reproduce:** Just press `'`. | non_defect | auto completion in comments not entirely useful for apostrophes godot version stable os device including version n a issue description auto completion has its issues when you re writing comments for example can t and won t do this would duplicate the character i e treat it as a single quote this is a minor issue but makes contractions i e it s can t etc slightly more difficult than it needs to be maybe look at the way other projects such as inform for example handle this steps to reproduce just press | 0 |
162,523 | 13,890,776,920 | IssuesEvent | 2020-10-19 09:44:05 | iearn-finance/docs | https://api.github.com/repos/iearn-finance/docs | closed | yInsure/yCover docs out of date | documentation enhancement good first issue help wanted | Recently the community discussed that "insurance" is inaccurate for describing "coverage" through Nexus Mutual and also has adverse regulatory implications.
The terminology has been changed from "insurance" to "cover" in some places on yearn.finance, but [yInsure docs](https://docs.yearn.finance/products/yinsure) should be further updated to make the terminology change consistent.
Also suggest adding a link on the same page to Nexus Mutual materials (or other materials, if Nexus Mutual has not published them) explaining the difference between coverage and insurance. | 1.0 | yInsure/yCover docs out of date - Recently the community discussed that "insurance" is inaccurate for describing "coverage" through Nexus Mutual and also has adverse regulatory implications.
The terminology has been changed from "insurance" to "cover" in some places on yearn.finance, but [yInsure docs](https://docs.yearn.finance/products/yinsure) should be further updated to make the terminology change consistent.
Also suggest adding a link on the same page to Nexus Mutual materials (or other materials, if Nexus Mutual has not published them) explaining the difference between coverage and insurance. | non_defect | yinsure ycover docs out of date recently the community discussed that insurance is inaccurate for describing coverage through nexus mutual and also has adverse regulatory implications the terminology has been changed from insurance to cover in some places on yearn finance but should be further updated to make the terminology change consistent also suggest adding a link on the same page to nexus mutual materials or other materials if nexus mutual has not published them explaining the difference between coverage and insurance | 0 |
26,919 | 6,812,765,661 | IssuesEvent | 2017-11-06 05:39:00 | BTDF/DeploymentFramework | https://api.github.com/repos/BTDF/DeploymentFramework | closed | Add support for BizTalk 2009 | CodePlexMigrationInitiated enhancement General Impact: Medium Release 5.0 | Add support for BizTalk 2009
#### This work item was migrated from CodePlex
CodePlex work item ID: '4060'
Assigned to: 'tfabraham'
Vote count: '1'
| 1.0 | Add support for BizTalk 2009 - Add support for BizTalk 2009
#### This work item was migrated from CodePlex
CodePlex work item ID: '4060'
Assigned to: 'tfabraham'
Vote count: '1'
| non_defect | add support for biztalk add support for biztalk this work item was migrated from codeplex codeplex work item id assigned to tfabraham vote count | 0 |
69,931 | 17,929,137,499 | IssuesEvent | 2021-09-10 06:42:21 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Hwloc mirror - download issue | stat:awaiting response stat:awaiting tensorflower type:build/install stalled subtype: raspberry pi | I am trying to cross-compile TF for Raspberry Pi via Docker container - as described in documentation. Unfortunately. it breaks with the following message:
ERROR: /workspace/tensorflow/core/BUILD:2432:1: no such package '@hwloc//': java.io.IOException: Error downloading [http://mirror.tensorflow.org/download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz, https://download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz] to /home/dmitry/tensorflow/bazel-ci_build-cache/.cache/bazel/_bazel_dmitry/eab0d61a99b6696edb3d2aff87b585e8/external/hwloc/hwloc-2.0.3.tar.gz: Tried to reconnect at offset 6,391,630 but server didn't support it and referenced by '//tensorflow/core:lib_internal_impl'
ERROR: Analysis of target '//tensorflow/tools/pip_package:build_pip_package' failed; build aborted: no such package '@hwloc//': java.io.IOException: Error downloading [http://mirror.tensorflow.org/download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz, https://download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz] to /home/dmitry/tensorflow/bazel-ci_build-cache/.cache/bazel/_bazel_dmitry/eab0d61a99b6696edb3d2aff87b585e8/external/hwloc/hwloc-2.0.3.tar.gz: Tried to reconnect at offset 6,391,630 but server didn't support it
Fixes tried with no luck:
- remove a semi-downloaded package from Bazel cache
- download it from the original site and put to a cache
- download from a mirror with Chrome (breaks in two parts)
**System information**
TF_BUILD_INFO = {
container_type: "pi-python3",
command: "tensorflow/tools/ci_build/pi/build_raspberry_pi.sh PI_ONE",
source_HEAD: "c407b045b8802f9eded430ef48be18cd85e4788c",
source_remote_origin: "https://github.com/tensorflow/tensorflow.git",
OS: "Linux",
kernel: "4.15.0-54-generic",
architecture: "x86_64",
processor: "Intel(R) Core(TM) i7-8550U CPU @ 1.80GHz",
processor_count: "8",
memory_total: "16305540 kB",
swap_total: "16657404 kB",
Bazel_version: "Build label: 0.24.1",
Java_version: "1.8.0_222-ea",
Python_version: "2.7.6",
gpp_version: "g++ (Ubuntu 4.8.4-2ubuntu1~14.04.4) 4.8.4",
swig_version: "",
NVIDIA_driver_version: "418.56",
CUDA_device_count: "0",
CUDA_device_names: "",
CUDA_toolkit_version: ""
}
**Provide the exact sequence of commands / steps that you executed before running into the problem**
CI_DOCKER_EXTRA_PARAMS="-e CI_BUILD_PYTHON=python3 -e CROSSTOOL_PYTHON_INCLUDE_PATH=/usr/include/python3.4" tensorflow/tools/ci_build/ci_build.sh PI-PYTHON3 tensorflow/tools/ci_build/pi/build_raspberry_pi.sh PI_ONE
| 1.0 | Hwloc mirror - download issue - I am trying to cross-compile TF for Raspberry Pi via Docker container - as described in documentation. Unfortunately. it breaks with the following message:
ERROR: /workspace/tensorflow/core/BUILD:2432:1: no such package '@hwloc//': java.io.IOException: Error downloading [http://mirror.tensorflow.org/download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz, https://download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz] to /home/dmitry/tensorflow/bazel-ci_build-cache/.cache/bazel/_bazel_dmitry/eab0d61a99b6696edb3d2aff87b585e8/external/hwloc/hwloc-2.0.3.tar.gz: Tried to reconnect at offset 6,391,630 but server didn't support it and referenced by '//tensorflow/core:lib_internal_impl'
ERROR: Analysis of target '//tensorflow/tools/pip_package:build_pip_package' failed; build aborted: no such package '@hwloc//': java.io.IOException: Error downloading [http://mirror.tensorflow.org/download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz, https://download.open-mpi.org/release/hwloc/v2.0/hwloc-2.0.3.tar.gz] to /home/dmitry/tensorflow/bazel-ci_build-cache/.cache/bazel/_bazel_dmitry/eab0d61a99b6696edb3d2aff87b585e8/external/hwloc/hwloc-2.0.3.tar.gz: Tried to reconnect at offset 6,391,630 but server didn't support it
Fixes tried with no luck:
- remove a semi-downloaded package from Bazel cache
- download it from the original site and put to a cache
- download from a mirror with Chrome (breaks in two parts)
**System information**
TF_BUILD_INFO = {
container_type: "pi-python3",
command: "tensorflow/tools/ci_build/pi/build_raspberry_pi.sh PI_ONE",
source_HEAD: "c407b045b8802f9eded430ef48be18cd85e4788c",
source_remote_origin: "https://github.com/tensorflow/tensorflow.git",
OS: "Linux",
kernel: "4.15.0-54-generic",
architecture: "x86_64",
processor: "Intel(R) Core(TM) i7-8550U CPU @ 1.80GHz",
processor_count: "8",
memory_total: "16305540 kB",
swap_total: "16657404 kB",
Bazel_version: "Build label: 0.24.1",
Java_version: "1.8.0_222-ea",
Python_version: "2.7.6",
gpp_version: "g++ (Ubuntu 4.8.4-2ubuntu1~14.04.4) 4.8.4",
swig_version: "",
NVIDIA_driver_version: "418.56",
CUDA_device_count: "0",
CUDA_device_names: "",
CUDA_toolkit_version: ""
}
**Provide the exact sequence of commands / steps that you executed before running into the problem**
CI_DOCKER_EXTRA_PARAMS="-e CI_BUILD_PYTHON=python3 -e CROSSTOOL_PYTHON_INCLUDE_PATH=/usr/include/python3.4" tensorflow/tools/ci_build/ci_build.sh PI-PYTHON3 tensorflow/tools/ci_build/pi/build_raspberry_pi.sh PI_ONE
| non_defect | hwloc mirror download issue i am trying to cross compile tf for raspberry pi via docker container as described in documentation unfortunately it breaks with the following message error workspace tensorflow core build no such package hwloc java io ioexception error downloading to home dmitry tensorflow bazel ci build cache cache bazel bazel dmitry external hwloc hwloc tar gz tried to reconnect at offset but server didn t support it and referenced by tensorflow core lib internal impl error analysis of target tensorflow tools pip package build pip package failed build aborted no such package hwloc java io ioexception error downloading to home dmitry tensorflow bazel ci build cache cache bazel bazel dmitry external hwloc hwloc tar gz tried to reconnect at offset but server didn t support it fixes tried with no luck remove a semi downloaded package from bazel cache download it from the original site and put to a cache download from a mirror with chrome breaks in two parts system information tf build info container type pi command tensorflow tools ci build pi build raspberry pi sh pi one source head source remote origin os linux kernel generic architecture processor intel r core tm cpu processor count memory total kb swap total kb bazel version build label java version ea python version gpp version g ubuntu swig version nvidia driver version cuda device count cuda device names cuda toolkit version provide the exact sequence of commands steps that you executed before running into the problem ci docker extra params e ci build python e crosstool python include path usr include tensorflow tools ci build ci build sh pi tensorflow tools ci build pi build raspberry pi sh pi one | 0 |
18,409 | 3,058,092,170 | IssuesEvent | 2015-08-14 04:16:52 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | Endles Request with Digest Authentication in SSL Context | Defect httpsocket On hold | I use the current stable Release of CakePhp (2.4.2-0-g75624c0).
If I want to connect a digest secured side with Cake\Network\Http\HttpSocket I get an endles request which ends after 100 times.
I searched and found a solution for this problem. If you swap lines 332 and 333, the problem don't occur. This is because the realm and nonce are the before the next request begins:
This is the right order:
$this->request['auth'] = $this->_auth;
$this->_setAuth(); | 1.0 | Endles Request with Digest Authentication in SSL Context - I use the current stable Release of CakePhp (2.4.2-0-g75624c0).
If I want to connect a digest secured side with Cake\Network\Http\HttpSocket I get an endles request which ends after 100 times.
I searched and found a solution for this problem. If you swap lines 332 and 333, the problem don't occur. This is because the realm and nonce are the before the next request begins:
This is the right order:
$this->request['auth'] = $this->_auth;
$this->_setAuth(); | defect | endles request with digest authentication in ssl context i use the current stable release of cakephp if i want to connect a digest secured side with cake network http httpsocket i get an endles request which ends after times i searched and found a solution for this problem if you swap lines and the problem don t occur this is because the realm and nonce are the before the next request begins this is the right order this request this auth this setauth | 1 |
35,737 | 7,800,296,924 | IssuesEvent | 2018-06-09 07:37:42 | StrikeNP/trac_test | https://api.github.com/repos/StrikeNP/trac_test | closed | CLUBB does not compile with -DNETCDF turned off (Trac #515) | Migrated from Trac clubb_src defect dschanen@uwm.edu | CLUBB does not compile with the DNETCDF preprocessor flag turned off. The error is found in input_fields.F90:
```text
../src/input_fields.F90:2521.61:
write(fstderr,*) "Error reading file "// trim( filename )
1
Error: Symbol 'filename' at (1) has no IMPLICIT type
```
Additionally, there are a number of new compiler warnings when DNETCDF is turned off.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/515
```json
{
"status": "closed",
"changetime": "2012-07-19T19:10:59",
"description": "CLUBB does not compile with the DNETCDF preprocessor flag turned off. The error is found in input_fields.F90:\n{{{\n../src/input_fields.F90:2521.61:\n\n write(fstderr,*) \"Error reading file \"// trim( filename )\n 1\nError: Symbol 'filename' at (1) has no IMPLICIT type\n}}}\nAdditionally, there are a number of new compiler warnings when DNETCDF is turned off. \n",
"reporter": "bmg2@uwm.edu",
"cc": "vlarson@uwm.edu, bladornr@uwm.edu",
"resolution": "fixed",
"_ts": "1342725059818255",
"component": "clubb_src",
"summary": "CLUBB does not compile with -DNETCDF turned off",
"priority": "minor",
"keywords": "Preprocessor flags",
"time": "2012-07-19T06:35:11",
"milestone": "",
"owner": "dschanen@uwm.edu",
"type": "defect"
}
```
| 1.0 | CLUBB does not compile with -DNETCDF turned off (Trac #515) - CLUBB does not compile with the DNETCDF preprocessor flag turned off. The error is found in input_fields.F90:
```text
../src/input_fields.F90:2521.61:
write(fstderr,*) "Error reading file "// trim( filename )
1
Error: Symbol 'filename' at (1) has no IMPLICIT type
```
Additionally, there are a number of new compiler warnings when DNETCDF is turned off.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/515
```json
{
"status": "closed",
"changetime": "2012-07-19T19:10:59",
"description": "CLUBB does not compile with the DNETCDF preprocessor flag turned off. The error is found in input_fields.F90:\n{{{\n../src/input_fields.F90:2521.61:\n\n write(fstderr,*) \"Error reading file \"// trim( filename )\n 1\nError: Symbol 'filename' at (1) has no IMPLICIT type\n}}}\nAdditionally, there are a number of new compiler warnings when DNETCDF is turned off. \n",
"reporter": "bmg2@uwm.edu",
"cc": "vlarson@uwm.edu, bladornr@uwm.edu",
"resolution": "fixed",
"_ts": "1342725059818255",
"component": "clubb_src",
"summary": "CLUBB does not compile with -DNETCDF turned off",
"priority": "minor",
"keywords": "Preprocessor flags",
"time": "2012-07-19T06:35:11",
"milestone": "",
"owner": "dschanen@uwm.edu",
"type": "defect"
}
```
| defect | clubb does not compile with dnetcdf turned off trac clubb does not compile with the dnetcdf preprocessor flag turned off the error is found in input fields text src input fields write fstderr error reading file trim filename error symbol filename at has no implicit type additionally there are a number of new compiler warnings when dnetcdf is turned off attachments migrated from json status closed changetime description clubb does not compile with the dnetcdf preprocessor flag turned off the error is found in input fields n n src input fields n n write fstderr error reading file trim filename n nerror symbol filename at has no implicit type n nadditionally there are a number of new compiler warnings when dnetcdf is turned off n reporter uwm edu cc vlarson uwm edu bladornr uwm edu resolution fixed ts component clubb src summary clubb does not compile with dnetcdf turned off priority minor keywords preprocessor flags time milestone owner dschanen uwm edu type defect | 1 |
168,886 | 26,709,710,886 | IssuesEvent | 2023-01-27 22:00:51 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Posibility to Customize Model Binding errors Key | :heavy_check_mark: Resolution: By Design feature-Model-Binding Status: Resolved area-web-frameworks | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe the problem.
I am using SuppressInferBindingSourcesForParameters for all the EP of an API.
And all the request that have a Body are defined this way:
```
[HttpPost("{Id}/MyMethod"]
public async Task<IActionResult> MyMethod(MyMethodRequest myMethodRequest)
```
The class is defined here:
```
public class MyMethodRequest
{
[FromRoute]
public int Id { get; set; }
[FromBody]
public Body BodyRequest { get; set; }
}
public class BodyRequest
{
public bool? Default { get; set; }
}
```
The response that ASP.NET when passing in the Body a Default that is not a bool is:
```
{
"errors": {
"body.Default": [
"Could not convert string to boolean: *. Path 'Default', line 8, position 19."
]
},
"type": "https://tools.ietf.org/html/rfc7231#section-6.5.1",
"title": "One or more validation errors occurred.",
"status": 400,
"traceId": "|7da03779-4ff981248d9543e6."
}
```
I have FluentValidation also applied to all the request and I am using OverridePropertyName to take out the Body part.
In FluentValidation for these cases we have to deal with Model Binding errors, the Body is null for this scenario and we have to override the PreValidate.
### Describe the solution you'd like
I just want that the response that ASP.Net gives back would be
```
{
"errors": {
"Default": [
"Could not convert string to boolean: *. Path 'Default', line 8, position 19."
]
},
"type": "https://tools.ietf.org/html/rfc7231#section-6.5.1",
"title": "One or more validation errors occurred.",
"status": 400,
"traceId": "|7da03779-4ff981248d9543e6."
}
```
### Additional context
_No response_ | 1.0 | Posibility to Customize Model Binding errors Key - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Is your feature request related to a problem? Please describe the problem.
I am using SuppressInferBindingSourcesForParameters for all the EP of an API.
And all the request that have a Body are defined this way:
```
[HttpPost("{Id}/MyMethod"]
public async Task<IActionResult> MyMethod(MyMethodRequest myMethodRequest)
```
The class is defined here:
```
public class MyMethodRequest
{
[FromRoute]
public int Id { get; set; }
[FromBody]
public Body BodyRequest { get; set; }
}
public class BodyRequest
{
public bool? Default { get; set; }
}
```
The response that ASP.NET when passing in the Body a Default that is not a bool is:
```
{
"errors": {
"body.Default": [
"Could not convert string to boolean: *. Path 'Default', line 8, position 19."
]
},
"type": "https://tools.ietf.org/html/rfc7231#section-6.5.1",
"title": "One or more validation errors occurred.",
"status": 400,
"traceId": "|7da03779-4ff981248d9543e6."
}
```
I have FluentValidation also applied to all the request and I am using OverridePropertyName to take out the Body part.
In FluentValidation for these cases we have to deal with Model Binding errors, the Body is null for this scenario and we have to override the PreValidate.
### Describe the solution you'd like
I just want that the response that ASP.Net gives back would be
```
{
"errors": {
"Default": [
"Could not convert string to boolean: *. Path 'Default', line 8, position 19."
]
},
"type": "https://tools.ietf.org/html/rfc7231#section-6.5.1",
"title": "One or more validation errors occurred.",
"status": 400,
"traceId": "|7da03779-4ff981248d9543e6."
}
```
### Additional context
_No response_ | non_defect | posibility to customize model binding errors key is there an existing issue for this i have searched the existing issues is your feature request related to a problem please describe the problem i am using suppressinferbindingsourcesforparameters for all the ep of an api and all the request that have a body are defined this way public async task mymethod mymethodrequest mymethodrequest the class is defined here public class mymethodrequest public int id get set public body bodyrequest get set public class bodyrequest public bool default get set the response that asp net when passing in the body a default that is not a bool is errors body default could not convert string to boolean path default line position type title one or more validation errors occurred status traceid i have fluentvalidation also applied to all the request and i am using overridepropertyname to take out the body part in fluentvalidation for these cases we have to deal with model binding errors the body is null for this scenario and we have to override the prevalidate describe the solution you d like i just want that the response that asp net gives back would be errors default could not convert string to boolean path default line position type title one or more validation errors occurred status traceid additional context no response | 0 |
43,006 | 11,426,946,977 | IssuesEvent | 2020-02-03 23:12:59 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Client is not Active using docker-compose AWS | Type: Defect | I'm seeing the following error messages when attempting to communicate with hazelcast:
[DefaultLogger] INFO at LifecycleService: HazelcastClient is starting
DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
....
after 5 minutes I see the following:
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
SEVERE: [hazelcast-local]:5701 [dev] [3.12.5] Could not join cluster. Shutting down now!
Jan 29, 2020 5:20:50 PM com.hazelcast.core.LifecycleService
INFO: [hazelcast-local]:5701 [dev] [3.12.5] [hazelcast-local]:5701 is SHUTTING_DOWN
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
WARNING: [hazelcast-local]:5701 [dev] [3.12.5] Terminating forcefully...
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Shutting down connection manager...
Jan 29, 2020 5:20:50 PM com.hazelcast.nio.tcp.TcpIpConnection
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connection[id=2, /172.23.0.4:46041->/172.31.13.13:5701, qualifier=null, endpoint=[172.31.13.13]:5701, alive=false, type=MEMBER] closed. Reason: EndpointManager is stopping
Jan 29, 2020 5:20:50 PM com.hazelcast.nio.tcp.TcpIpConnection
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connection[id=1, /172.23.0.4:5701->/172.23.0.1:41806, qualifier=null, endpoint=[hazelcast-local]:5701, alive=false, type=MEMBER] closed. Reason: Connection closed by the other side
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Shutting down node engine...
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.NodeExtension
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Destroying node NodeExtension.
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Hazelcast Shutdown is completed in 23 ms.
Jan 29, 2020 5:20:50 PM com.hazelcast.core.LifecycleService
INFO: [hazelcast-local]:5701 [dev] [3.12.5] [hazelcast-local]:5701 is SHUTDOWN
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
WARNING: [hazelcast-local]:5701 [dev] [3.12.5] ManagementCenterService could not be constructed!
com.hazelcast.spi.exception.RetryableHazelcastException: HazelcastInstance[[hazelcast-local]:5701] is not active!
at com.hazelcast.spi.impl.NodeEngineImpl.getService(NodeEngineImpl.java:370)
at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.getSegment(EventServiceImpl.java:543)
at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.registerListenerInternal(EventServiceImpl.java:284)
at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.registerLocalListener(EventServiceImpl.java:244)
at com.hazelcast.internal.cluster.impl.ClusterServiceImpl.addMembershipListener(ClusterServiceImpl.java:768)
at com.hazelcast.internal.management.ManagementCenterService.<init>(ManagementCenterService.java:144)
at com.hazelcast.instance.Node.start(Node.java:452)
at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:136)
at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:228)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:207)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:157)
at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:91)
at com.hazelcast.core.server.StartServer.main(StartServer.java:46)
Jan 29, 2020 5:20:50 PM com.hazelcast.util.PhoneHome
WARNING: [hazelcast-local]:5701 [dev] [3.12.5] Could not schedule phone home task! Most probably Hazelcast failed to start.
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Node is already shutting down... Waiting for shutdown process to complete...
Exception in thread "main" java.lang.IllegalStateException: Node failed to start!
at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:138)
at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:228)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:207)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:157)
at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:91)
at com.hazelcast.core.server.StartServer.main(StartServer.java:46)
The basic config within my hazelcast.xml is:
`
<network>
<port auto-increment="true" port-count="100">5701</port>
<outbound-ports>
<!--
Allowed port range when connecting to other nodes.
0 or * means use system provided port.
-->
<ports>0</ports>
</outbound-ports>
<join>
...
<aws enabled="true">
<iam-role>data-infrastructure-ums</iam-role>
<region>us-east-1</region>
<host-header>ec2.amazonaws.com</host-header>
<security-group-name>UMS HAZELCAST</security-group-name>
</aws>
...
<discovery-strategies>
</discovery-strategies>
</join>
<interfaces enabled="true">
<interface>172.*.*.*</interface>
</interfaces>
<ssl enabled="false"/>
`
If I toggle to aws enabled to false it works fine. Key take away is I'm using docker compose with all the containers running in the same instance. I also set JAVA_OPTS=-Dhazelcast.config=/opt/hazelcast/config_ext/hazelcast.xml -Dhazelcast.local.publicAddress=hazelcast-local:5701 in the docker compose file and seems to work as expected based on logs below. Is there something else I'm missing? I even tried to enable interfaces to see if that would solve it based on https://stackoverflow.com/questions/47302340/hazelcast-cluster-over-aws-using-docker. I would truly love for somebody to point me in the right direction.
I'm expected Hazelcast to work fine without issues as it seems to connect fine. Here are the log messages during connection:
hazelcast-local | ########################################
hazelcast-local | # JAVA_OPTS=-Dhazelcast.mancenter.enabled=false -Xmx512M -Xms256M -Djava.net.preferIPv4Stack=true -Djava.util.logging.config.file=/opt/hazelcast/logging.properties -Dhazelcast.config=/opt/hazelcast/config_ext/hazelcast.xml -Dhazelcast.local.publicAddress=hazelcast-local:5701 -Dhazelcast.diagnostics.enabled=true
hazelcast-local | # CLASSPATH=/opt/hazelcast/*:/opt/hazelcast/lib/*
hazelcast-local | # starting now....
hazelcast-local | ########################################
hazelcast-local | + exec java -server -Dhazelcast.mancenter.enabled=false -Xmx512M -Xms256M -Djava.net.preferIPv4Stack=true -Djava.util.logging.config.file=/opt/hazelcast/logging.properties -Dhazelcast.config=/opt/hazelcast/config_ext/hazelcast.xml -Dhazelcast.local.publicAddress=hazelcast-local:5701 -Dhazelcast.diagnostics.enabled=true com.hazelcast.core.server.StartServer
hazelcast-local | Jan 29, 2020 5:15:47 PM com.hazelcast.config.AbstractConfigLocator
hazelcast-local | INFO: Loading configuration '/opt/hazelcast/config_ext/hazelcast.xml' from System property 'hazelcast.config'
hazelcast-local | Jan 29, 2020 5:15:47 PM com.hazelcast.config.AbstractConfigLocator
hazelcast-local | INFO: Using configuration file at /opt/hazelcast/config_ext/hazelcast.xml
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Interfaces is enabled, trying to pick one address matching to one of: [172.*.*.*]
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Prefer IPv4 stack is true, prefer IPv6 addresses is false
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Picked [172.23.0.4]:5701, using socket ServerSocket[addr=/0.0.0.0,localport=5701], bind any local is true
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Using public address: [hazelcast-local]:5701
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.system
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Hazelcast 3.12.5 (20191210 - 294ff46) starting at [hazelcast-local]:5701
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.system
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Copyright (c) 2008-2019, Hazelcast, Inc. All Rights Reserved.
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.spi.impl.operationservice.impl.BackpressureRegulator
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Backpressure is disabled
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.aws.AwsDiscoveryStrategy
hazelcast-local | INFO: Describe instances will be queried with iam-role, please make sure given iam-role have ec2:DescribeInstances policy attached.
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.Node
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Activating Discovery SPI Joiner
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Starting 2 partition threads and 3 generic threads (1 dedicated for priority tasks)
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.Diagnostics
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Diagnostics started
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.BuildInfoPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.SystemPropertiesPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.ConfigPropertiesPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.MetricsPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:60000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.SlowOperationPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:60000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.MemberHazelcastInstanceInfoPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:60000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.SystemLogPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active: logPartitions:false
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.MemberHazelcastInstanceInfoPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:10000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.OperationHeartbeatPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active: period-millis:10000 max-deviation:33%
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.core.LifecycleService
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [hazelcast-local]:5701 is STARTING
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5704, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5706, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5701, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5705, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5706. Reason: SocketException[Connection refused to address /172.31.13.13:5706]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5706 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5708, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5708. Reason: SocketException[Connection refused to address /172.31.13.13:5708]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5708 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5702, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5702. Reason: SocketException[Connection refused to address /172.31.13.13:5702]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5702 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5704. Reason: SocketException[Connection refused to address /172.31.13.13:5704]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5704 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5707, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5707. Reason: SocketException[Connection refused to address /172.31.13.13:5707]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5707 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5703, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5703. Reason: SocketException[Connection refused to address /172.31.13.13:5703]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5703 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5705. Reason: SocketException[Connection refused to address /172.31.13.13:5705]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnection
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Initialized new cluster connection between /172.23.0.4:5701 and /172.23.0.1:41806
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5705 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnection
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Initialized new cluster connection between /172.23.0.4:46041 and /172.31.13.13:5701
I'm using the hazelcast docker image 3.12.5
The AWS instance I'm using is c5 with 2 cpus and 4 gb. Also I have the security group and IAM role setup correctly. I only have one cluster member.
| 1.0 | Client is not Active using docker-compose AWS - I'm seeing the following error messages when attempting to communicate with hazelcast:
[DefaultLogger] INFO at LifecycleService: HazelcastClient is starting
DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
[DefaultLogger] WARN at ClusterService: Error: Client is not active.
....
after 5 minutes I see the following:
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
SEVERE: [hazelcast-local]:5701 [dev] [3.12.5] Could not join cluster. Shutting down now!
Jan 29, 2020 5:20:50 PM com.hazelcast.core.LifecycleService
INFO: [hazelcast-local]:5701 [dev] [3.12.5] [hazelcast-local]:5701 is SHUTTING_DOWN
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
WARNING: [hazelcast-local]:5701 [dev] [3.12.5] Terminating forcefully...
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Shutting down connection manager...
Jan 29, 2020 5:20:50 PM com.hazelcast.nio.tcp.TcpIpConnection
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connection[id=2, /172.23.0.4:46041->/172.31.13.13:5701, qualifier=null, endpoint=[172.31.13.13]:5701, alive=false, type=MEMBER] closed. Reason: EndpointManager is stopping
Jan 29, 2020 5:20:50 PM com.hazelcast.nio.tcp.TcpIpConnection
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connection[id=1, /172.23.0.4:5701->/172.23.0.1:41806, qualifier=null, endpoint=[hazelcast-local]:5701, alive=false, type=MEMBER] closed. Reason: Connection closed by the other side
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Shutting down node engine...
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.NodeExtension
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Destroying node NodeExtension.
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Hazelcast Shutdown is completed in 23 ms.
Jan 29, 2020 5:20:50 PM com.hazelcast.core.LifecycleService
INFO: [hazelcast-local]:5701 [dev] [3.12.5] [hazelcast-local]:5701 is SHUTDOWN
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
WARNING: [hazelcast-local]:5701 [dev] [3.12.5] ManagementCenterService could not be constructed!
com.hazelcast.spi.exception.RetryableHazelcastException: HazelcastInstance[[hazelcast-local]:5701] is not active!
at com.hazelcast.spi.impl.NodeEngineImpl.getService(NodeEngineImpl.java:370)
at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.getSegment(EventServiceImpl.java:543)
at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.registerListenerInternal(EventServiceImpl.java:284)
at com.hazelcast.spi.impl.eventservice.impl.EventServiceImpl.registerLocalListener(EventServiceImpl.java:244)
at com.hazelcast.internal.cluster.impl.ClusterServiceImpl.addMembershipListener(ClusterServiceImpl.java:768)
at com.hazelcast.internal.management.ManagementCenterService.<init>(ManagementCenterService.java:144)
at com.hazelcast.instance.Node.start(Node.java:452)
at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:136)
at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:228)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:207)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:157)
at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:91)
at com.hazelcast.core.server.StartServer.main(StartServer.java:46)
Jan 29, 2020 5:20:50 PM com.hazelcast.util.PhoneHome
WARNING: [hazelcast-local]:5701 [dev] [3.12.5] Could not schedule phone home task! Most probably Hazelcast failed to start.
Jan 29, 2020 5:20:50 PM com.hazelcast.instance.Node
INFO: [hazelcast-local]:5701 [dev] [3.12.5] Node is already shutting down... Waiting for shutdown process to complete...
Exception in thread "main" java.lang.IllegalStateException: Node failed to start!
at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:138)
at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:228)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:207)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:157)
at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:91)
at com.hazelcast.core.server.StartServer.main(StartServer.java:46)
The basic config within my hazelcast.xml is:
`
<network>
<port auto-increment="true" port-count="100">5701</port>
<outbound-ports>
<!--
Allowed port range when connecting to other nodes.
0 or * means use system provided port.
-->
<ports>0</ports>
</outbound-ports>
<join>
...
<aws enabled="true">
<iam-role>data-infrastructure-ums</iam-role>
<region>us-east-1</region>
<host-header>ec2.amazonaws.com</host-header>
<security-group-name>UMS HAZELCAST</security-group-name>
</aws>
...
<discovery-strategies>
</discovery-strategies>
</join>
<interfaces enabled="true">
<interface>172.*.*.*</interface>
</interfaces>
<ssl enabled="false"/>
`
If I toggle to aws enabled to false it works fine. Key take away is I'm using docker compose with all the containers running in the same instance. I also set JAVA_OPTS=-Dhazelcast.config=/opt/hazelcast/config_ext/hazelcast.xml -Dhazelcast.local.publicAddress=hazelcast-local:5701 in the docker compose file and seems to work as expected based on logs below. Is there something else I'm missing? I even tried to enable interfaces to see if that would solve it based on https://stackoverflow.com/questions/47302340/hazelcast-cluster-over-aws-using-docker. I would truly love for somebody to point me in the right direction.
I'm expected Hazelcast to work fine without issues as it seems to connect fine. Here are the log messages during connection:
hazelcast-local | ########################################
hazelcast-local | # JAVA_OPTS=-Dhazelcast.mancenter.enabled=false -Xmx512M -Xms256M -Djava.net.preferIPv4Stack=true -Djava.util.logging.config.file=/opt/hazelcast/logging.properties -Dhazelcast.config=/opt/hazelcast/config_ext/hazelcast.xml -Dhazelcast.local.publicAddress=hazelcast-local:5701 -Dhazelcast.diagnostics.enabled=true
hazelcast-local | # CLASSPATH=/opt/hazelcast/*:/opt/hazelcast/lib/*
hazelcast-local | # starting now....
hazelcast-local | ########################################
hazelcast-local | + exec java -server -Dhazelcast.mancenter.enabled=false -Xmx512M -Xms256M -Djava.net.preferIPv4Stack=true -Djava.util.logging.config.file=/opt/hazelcast/logging.properties -Dhazelcast.config=/opt/hazelcast/config_ext/hazelcast.xml -Dhazelcast.local.publicAddress=hazelcast-local:5701 -Dhazelcast.diagnostics.enabled=true com.hazelcast.core.server.StartServer
hazelcast-local | Jan 29, 2020 5:15:47 PM com.hazelcast.config.AbstractConfigLocator
hazelcast-local | INFO: Loading configuration '/opt/hazelcast/config_ext/hazelcast.xml' from System property 'hazelcast.config'
hazelcast-local | Jan 29, 2020 5:15:47 PM com.hazelcast.config.AbstractConfigLocator
hazelcast-local | INFO: Using configuration file at /opt/hazelcast/config_ext/hazelcast.xml
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Interfaces is enabled, trying to pick one address matching to one of: [172.*.*.*]
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Prefer IPv4 stack is true, prefer IPv6 addresses is false
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Picked [172.23.0.4]:5701, using socket ServerSocket[addr=/0.0.0.0,localport=5701], bind any local is true
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.AddressPicker
hazelcast-local | INFO: [LOCAL] [dev] [3.12.5] Using public address: [hazelcast-local]:5701
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.system
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Hazelcast 3.12.5 (20191210 - 294ff46) starting at [hazelcast-local]:5701
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.system
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Copyright (c) 2008-2019, Hazelcast, Inc. All Rights Reserved.
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.spi.impl.operationservice.impl.BackpressureRegulator
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Backpressure is disabled
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.aws.AwsDiscoveryStrategy
hazelcast-local | INFO: Describe instances will be queried with iam-role, please make sure given iam-role have ec2:DescribeInstances policy attached.
hazelcast-local | Jan 29, 2020 5:15:48 PM com.hazelcast.instance.Node
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Activating Discovery SPI Joiner
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Starting 2 partition threads and 3 generic threads (1 dedicated for priority tasks)
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.Diagnostics
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Diagnostics started
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.BuildInfoPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.SystemPropertiesPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.ConfigPropertiesPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.MetricsPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:60000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.SlowOperationPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:60000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.MemberHazelcastInstanceInfoPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:60000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.SystemLogPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active: logPartitions:false
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.MemberHazelcastInstanceInfoPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active, period-millis:10000
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.internal.diagnostics.OperationHeartbeatPlugin
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Plugin:active: period-millis:10000 max-deviation:33%
hazelcast-local | Jan 29, 2020 5:15:49 PM com.hazelcast.core.LifecycleService
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [hazelcast-local]:5701 is STARTING
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5704, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5706, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5701, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5705, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5706. Reason: SocketException[Connection refused to address /172.31.13.13:5706]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5706 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5708, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5708. Reason: SocketException[Connection refused to address /172.31.13.13:5708]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5708 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5702, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5702. Reason: SocketException[Connection refused to address /172.31.13.13:5702]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5702 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5704. Reason: SocketException[Connection refused to address /172.31.13.13:5704]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5704 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5707, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5707. Reason: SocketException[Connection refused to address /172.31.13.13:5707]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5707 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Connecting to /172.31.13.13:5703, timeout: 10000, bind-any: true
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5703. Reason: SocketException[Connection refused to address /172.31.13.13:5703]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5703 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnector
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Could not connect to: /172.31.13.13:5705. Reason: SocketException[Connection refused to address /172.31.13.13:5705]
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnection
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Initialized new cluster connection between /172.23.0.4:5701 and /172.23.0.1:41806
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.internal.cluster.impl.DiscoveryJoiner
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] [172.31.13.13]:5705 is added to the blacklist.
hazelcast-local | Jan 29, 2020 5:15:50 PM com.hazelcast.nio.tcp.TcpIpConnection
hazelcast-local | INFO: [hazelcast-local]:5701 [dev] [3.12.5] Initialized new cluster connection between /172.23.0.4:46041 and /172.31.13.13:5701
I'm using the hazelcast docker image 3.12.5
The AWS instance I'm using is c5 with 2 cpus and 4 gb. Also I have the security group and IAM role setup correctly. I only have one cluster member.
| defect | client is not active using docker compose aws i m seeing the following error messages when attempting to communicate with hazelcast info at lifecycleservice hazelcastclient is starting defaultlogger warn at clusterservice error client is not active warn at clusterservice error client is not active warn at clusterservice error client is not active warn at clusterservice error client is not active warn at clusterservice error client is not active after minutes i see the following jan pm com hazelcast instance node severe could not join cluster shutting down now jan pm com hazelcast core lifecycleservice info is shutting down jan pm com hazelcast instance node warning terminating forcefully jan pm com hazelcast instance node info shutting down connection manager jan pm com hazelcast nio tcp tcpipconnection info connection alive false type member closed reason endpointmanager is stopping jan pm com hazelcast nio tcp tcpipconnection info connection alive false type member closed reason connection closed by the other side jan pm com hazelcast instance node info shutting down node engine jan pm com hazelcast instance nodeextension info destroying node nodeextension jan pm com hazelcast instance node info hazelcast shutdown is completed in ms jan pm com hazelcast core lifecycleservice info is shutdown jan pm com hazelcast instance node warning managementcenterservice could not be constructed com hazelcast spi exception retryablehazelcastexception hazelcastinstance is not active at com hazelcast spi impl nodeengineimpl getservice nodeengineimpl java at com hazelcast spi impl eventservice impl eventserviceimpl getsegment eventserviceimpl java at com hazelcast spi impl eventservice impl eventserviceimpl registerlistenerinternal eventserviceimpl java at com hazelcast spi impl eventservice impl eventserviceimpl registerlocallistener eventserviceimpl java at com hazelcast internal cluster impl clusterserviceimpl addmembershiplistener clusterserviceimpl java at com hazelcast internal management managementcenterservice managementcenterservice java at com hazelcast instance node start node java at com hazelcast instance hazelcastinstanceimpl hazelcastinstanceimpl java at com hazelcast instance hazelcastinstancefactory constructhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast core hazelcast newhazelcastinstance hazelcast java at com hazelcast core server startserver main startserver java jan pm com hazelcast util phonehome warning could not schedule phone home task most probably hazelcast failed to start jan pm com hazelcast instance node info node is already shutting down waiting for shutdown process to complete exception in thread main java lang illegalstateexception node failed to start at com hazelcast instance hazelcastinstanceimpl hazelcastinstanceimpl java at com hazelcast instance hazelcastinstancefactory constructhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast core hazelcast newhazelcastinstance hazelcast java at com hazelcast core server startserver main startserver java the basic config within my hazelcast xml is allowed port range when connecting to other nodes or means use system provided port data infrastructure ums us east amazonaws com ums hazelcast if i toggle to aws enabled to false it works fine key take away is i m using docker compose with all the containers running in the same instance i also set java opts dhazelcast config opt hazelcast config ext hazelcast xml dhazelcast local publicaddress hazelcast local in the docker compose file and seems to work as expected based on logs below is there something else i m missing i even tried to enable interfaces to see if that would solve it based on i would truly love for somebody to point me in the right direction i m expected hazelcast to work fine without issues as it seems to connect fine here are the log messages during connection hazelcast local hazelcast local java opts dhazelcast mancenter enabled false djava net true djava util logging config file opt hazelcast logging properties dhazelcast config opt hazelcast config ext hazelcast xml dhazelcast local publicaddress hazelcast local dhazelcast diagnostics enabled true hazelcast local classpath opt hazelcast opt hazelcast lib hazelcast local starting now hazelcast local hazelcast local exec java server dhazelcast mancenter enabled false djava net true djava util logging config file opt hazelcast logging properties dhazelcast config opt hazelcast config ext hazelcast xml dhazelcast local publicaddress hazelcast local dhazelcast diagnostics enabled true com hazelcast core server startserver hazelcast local jan pm com hazelcast config abstractconfiglocator hazelcast local info loading configuration opt hazelcast config ext hazelcast xml from system property hazelcast config hazelcast local jan pm com hazelcast config abstractconfiglocator hazelcast local info using configuration file at opt hazelcast config ext hazelcast xml hazelcast local jan pm com hazelcast instance addresspicker hazelcast local info interfaces is enabled trying to pick one address matching to one of hazelcast local jan pm com hazelcast instance addresspicker hazelcast local info prefer stack is true prefer addresses is false hazelcast local jan pm com hazelcast instance addresspicker hazelcast local info picked using socket serversocket bind any local is true hazelcast local jan pm com hazelcast instance addresspicker hazelcast local info using public address hazelcast local jan pm com hazelcast system hazelcast local info hazelcast starting at hazelcast local jan pm com hazelcast system hazelcast local info copyright c hazelcast inc all rights reserved hazelcast local jan pm com hazelcast spi impl operationservice impl backpressureregulator hazelcast local info backpressure is disabled hazelcast local jan pm com hazelcast aws awsdiscoverystrategy hazelcast local info describe instances will be queried with iam role please make sure given iam role have describeinstances policy attached hazelcast local jan pm com hazelcast instance node hazelcast local info activating discovery spi joiner hazelcast local jan pm com hazelcast spi impl operationexecutor impl operationexecutorimpl hazelcast local info starting partition threads and generic threads dedicated for priority tasks hazelcast local jan pm com hazelcast internal diagnostics diagnostics hazelcast local info diagnostics started hazelcast local jan pm com hazelcast internal diagnostics buildinfoplugin hazelcast local info plugin active hazelcast local jan pm com hazelcast internal diagnostics systempropertiesplugin hazelcast local info plugin active hazelcast local jan pm com hazelcast internal diagnostics configpropertiesplugin hazelcast local info plugin active hazelcast local jan pm com hazelcast internal diagnostics metricsplugin hazelcast local info plugin active period millis hazelcast local jan pm com hazelcast internal diagnostics slowoperationplugin hazelcast local info plugin active period millis hazelcast local jan pm com hazelcast internal diagnostics memberhazelcastinstanceinfoplugin hazelcast local info plugin active period millis hazelcast local jan pm com hazelcast internal diagnostics systemlogplugin hazelcast local info plugin active logpartitions false hazelcast local jan pm com hazelcast internal diagnostics memberhazelcastinstanceinfoplugin hazelcast local info plugin active period millis hazelcast local jan pm com hazelcast internal diagnostics operationheartbeatplugin hazelcast local info plugin active period millis max deviation hazelcast local jan pm com hazelcast core lifecycleservice hazelcast local info is starting hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info could not connect to reason socketexception hazelcast local jan pm com hazelcast internal cluster impl discoveryjoiner hazelcast local info is added to the blacklist hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info could not connect to reason socketexception hazelcast local jan pm com hazelcast internal cluster impl discoveryjoiner hazelcast local info is added to the blacklist hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info could not connect to reason socketexception hazelcast local jan pm com hazelcast internal cluster impl discoveryjoiner hazelcast local info is added to the blacklist hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info could not connect to reason socketexception hazelcast local jan pm com hazelcast internal cluster impl discoveryjoiner hazelcast local info is added to the blacklist hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info could not connect to reason socketexception hazelcast local jan pm com hazelcast internal cluster impl discoveryjoiner hazelcast local info is added to the blacklist hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info connecting to timeout bind any true hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info could not connect to reason socketexception hazelcast local jan pm com hazelcast internal cluster impl discoveryjoiner hazelcast local info is added to the blacklist hazelcast local jan pm com hazelcast nio tcp tcpipconnector hazelcast local info could not connect to reason socketexception hazelcast local jan pm com hazelcast nio tcp tcpipconnection hazelcast local info initialized new cluster connection between and hazelcast local jan pm com hazelcast internal cluster impl discoveryjoiner hazelcast local info is added to the blacklist hazelcast local jan pm com hazelcast nio tcp tcpipconnection hazelcast local info initialized new cluster connection between and i m using the hazelcast docker image the aws instance i m using is with cpus and gb also i have the security group and iam role setup correctly i only have one cluster member | 1 |
1,343 | 2,947,888,989 | IssuesEvent | 2015-07-05 16:04:58 | pouchdb/pouchdb | https://api.github.com/repos/pouchdb/pouchdb | closed | Improving replication performance, looking at the "_attachment fitting through XMLHttpRequest" bottleneck | feature performance | Firefox has always seemed to lock up quite dramatically when fitting just a few MB of data through XMLHttpRequest. Chrome is a little better in this department. It sure would be handy if there was something in the underlying data storage protocols that downloaded the file directly to disk without having to turn the file into a blob in memory before saving to disk. Does anyone know of anything like this? Anything even just experimental? If not something like just mentioned, does anyone have any ideas on improving performance of replication in other ways?
<bountysource-plugin>
---
Want to back this issue? **[Place a bounty on it!](https://www.bountysource.com/issues/796195-improving-replication-performance-looking-at-the-_attachment-fitting-through-xmlhttprequest-bottleneck?utm_campaign=plugin&utm_content=tracker%2F52197&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F52197&utm_medium=issues&utm_source=github).
</bountysource-plugin> | True | Improving replication performance, looking at the "_attachment fitting through XMLHttpRequest" bottleneck - Firefox has always seemed to lock up quite dramatically when fitting just a few MB of data through XMLHttpRequest. Chrome is a little better in this department. It sure would be handy if there was something in the underlying data storage protocols that downloaded the file directly to disk without having to turn the file into a blob in memory before saving to disk. Does anyone know of anything like this? Anything even just experimental? If not something like just mentioned, does anyone have any ideas on improving performance of replication in other ways?
<bountysource-plugin>
---
Want to back this issue? **[Place a bounty on it!](https://www.bountysource.com/issues/796195-improving-replication-performance-looking-at-the-_attachment-fitting-through-xmlhttprequest-bottleneck?utm_campaign=plugin&utm_content=tracker%2F52197&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F52197&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_defect | improving replication performance looking at the attachment fitting through xmlhttprequest bottleneck firefox has always seemed to lock up quite dramatically when fitting just a few mb of data through xmlhttprequest chrome is a little better in this department it sure would be handy if there was something in the underlying data storage protocols that downloaded the file directly to disk without having to turn the file into a blob in memory before saving to disk does anyone know of anything like this anything even just experimental if not something like just mentioned does anyone have any ideas on improving performance of replication in other ways want to back this issue we accept bounties via | 0 |
65,308 | 19,345,039,931 | IssuesEvent | 2021-12-15 09:56:17 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Room menu: same actions should have the same wording | T-Defect S-Minor A-Room-List Z-IA | |Top timeline menu|Menu from room list|
|-|-|
|<img width="473" alt="image" src="https://user-images.githubusercontent.com/3940906/145580670-42dee74d-ac51-47f5-a356-15329fc0da07.png">|<img width="467" alt="image" src="https://user-images.githubusercontent.com/3940906/145580728-50cc441b-b889-434b-93b9-8cb243a8edfb.png">|
See the diff for invite/room link/leave | 1.0 | Room menu: same actions should have the same wording - |Top timeline menu|Menu from room list|
|-|-|
|<img width="473" alt="image" src="https://user-images.githubusercontent.com/3940906/145580670-42dee74d-ac51-47f5-a356-15329fc0da07.png">|<img width="467" alt="image" src="https://user-images.githubusercontent.com/3940906/145580728-50cc441b-b889-434b-93b9-8cb243a8edfb.png">|
See the diff for invite/room link/leave | defect | room menu same actions should have the same wording top timeline menu menu from room list img width alt image src width alt image src see the diff for invite room link leave | 1 |
237,125 | 7,756,262,800 | IssuesEvent | 2018-05-31 13:06:36 | MARKETProtocol/MARKETProtocol | https://api.github.com/repos/MARKETProtocol/MARKETProtocol | closed | [npm] Rename the npm package to @marketprotocol/solidity | Priority: High Status: Pending Type: Enhancement | ### Description
Update the `npm` package to follow the Org scope of `@marketprotocol` for easy linking across all the other repos.
Ref: https://docs.npmjs.com/getting-started/scoped-packages
*Type*: Enhancement
### Definition of Done
- Change `market-solidity` to `@marketprotocol/solidity` | 1.0 | [npm] Rename the npm package to @marketprotocol/solidity - ### Description
Update the `npm` package to follow the Org scope of `@marketprotocol` for easy linking across all the other repos.
Ref: https://docs.npmjs.com/getting-started/scoped-packages
*Type*: Enhancement
### Definition of Done
- Change `market-solidity` to `@marketprotocol/solidity` | non_defect | rename the npm package to marketprotocol solidity description update the npm package to follow the org scope of marketprotocol for easy linking across all the other repos ref type enhancement definition of done change market solidity to marketprotocol solidity | 0 |
32,886 | 6,962,524,291 | IssuesEvent | 2017-12-08 14:05:51 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | Dialog state (maximized/minimized) buggy across AJAX updates | 6.1.11 defect | When a dialog is closed while in minimized/maximized state and then updated (via AJAX) its state become corrupted somehow.
If it was closed in maximized state, when it's shown again it's in normal size but it "thinks" it's still in maximized state so 2 clicks on the maximize button are required in order to maximize it.
If, on the other hand, it was closed while in minimized state, when it's shown again it appears behind the overlay (in case of overlay dialog) which makes the page unusable.
## 1) Environment
Tested on
Tomcat 7
Primfaces 6.1
Chrome
## 2) Expected behavior
The dialog should show in the former state, or at least be consistent (i.e. not show in regular state but thinking it's maximized and not blocking the page)
## 3) Actual behavior
- If the dialog was maximized before closing, two clicks on the maximize button are needed in order to maximize it
- If the dialog was minimized before closing and is modal, it's shown behind the overlay, making the page unusable
## 4) Steps to reproduce
1. Open a dialog
2. Maximize/minimize it
3. Close it
4. Fire any ajax that will cause the dialog to be updated
5. Open the dialog again
## 5) Sample XHTML
https://github.com/Adam-1/primefaces-test/blob/master/src/main/webapp/dialogTest.xhtml
## 6) Sample bean
N/A
| 1.0 | Dialog state (maximized/minimized) buggy across AJAX updates - When a dialog is closed while in minimized/maximized state and then updated (via AJAX) its state become corrupted somehow.
If it was closed in maximized state, when it's shown again it's in normal size but it "thinks" it's still in maximized state so 2 clicks on the maximize button are required in order to maximize it.
If, on the other hand, it was closed while in minimized state, when it's shown again it appears behind the overlay (in case of overlay dialog) which makes the page unusable.
## 1) Environment
Tested on
Tomcat 7
Primfaces 6.1
Chrome
## 2) Expected behavior
The dialog should show in the former state, or at least be consistent (i.e. not show in regular state but thinking it's maximized and not blocking the page)
## 3) Actual behavior
- If the dialog was maximized before closing, two clicks on the maximize button are needed in order to maximize it
- If the dialog was minimized before closing and is modal, it's shown behind the overlay, making the page unusable
## 4) Steps to reproduce
1. Open a dialog
2. Maximize/minimize it
3. Close it
4. Fire any ajax that will cause the dialog to be updated
5. Open the dialog again
## 5) Sample XHTML
https://github.com/Adam-1/primefaces-test/blob/master/src/main/webapp/dialogTest.xhtml
## 6) Sample bean
N/A
| defect | dialog state maximized minimized buggy across ajax updates when a dialog is closed while in minimized maximized state and then updated via ajax its state become corrupted somehow if it was closed in maximized state when it s shown again it s in normal size but it thinks it s still in maximized state so clicks on the maximize button are required in order to maximize it if on the other hand it was closed while in minimized state when it s shown again it appears behind the overlay in case of overlay dialog which makes the page unusable environment tested on tomcat primfaces chrome expected behavior the dialog should show in the former state or at least be consistent i e not show in regular state but thinking it s maximized and not blocking the page actual behavior if the dialog was maximized before closing two clicks on the maximize button are needed in order to maximize it if the dialog was minimized before closing and is modal it s shown behind the overlay making the page unusable steps to reproduce open a dialog maximize minimize it close it fire any ajax that will cause the dialog to be updated open the dialog again sample xhtml sample bean n a | 1 |
53,983 | 13,262,957,106 | IssuesEvent | 2020-08-20 23:05:17 | ophrescue/RescueRails | https://api.github.com/repos/ophrescue/RescueRails | closed | circular argument references in adopter_searcher | Defect | Warnings throwing during rspec run
```
app/models/adopter_searcher.rb:32: warning: circular argument reference - user_id
app/models/adopter_searcher.rb:68: warning: circular argument reference - user_id
```
Looks like this was introduced with 64a217f29d1dec28edd8ed5657898f819c2844b2 | 1.0 | circular argument references in adopter_searcher - Warnings throwing during rspec run
```
app/models/adopter_searcher.rb:32: warning: circular argument reference - user_id
app/models/adopter_searcher.rb:68: warning: circular argument reference - user_id
```
Looks like this was introduced with 64a217f29d1dec28edd8ed5657898f819c2844b2 | defect | circular argument references in adopter searcher warnings throwing during rspec run app models adopter searcher rb warning circular argument reference user id app models adopter searcher rb warning circular argument reference user id looks like this was introduced with | 1 |
308,423 | 9,439,052,654 | IssuesEvent | 2019-04-14 06:53:25 | CS2113-AY1819S2-T09-1/main | https://api.github.com/repos/CS2113-AY1819S2-T09-1/main | closed | As a user, I want to see specific degree planner(s). | priority.Low type.Enhancement type.Epic | The user may want to see the degree planners based on certain condition.
Hence, I propose implementing ```planner_show``` command as follows with/without Boolean expression:
- [x] able to show degree planner(s) by year
i.e. `planner_show year/YEAR`
returns academic semester(s) having its year matches the year given
- [x] able to show degree planner(s) by semester
i.e. `planner_show sem/SEMESTER`
returns academic semester(s) having its semester matches the semester given
- [x] able to include both `year` and `semester` attributes in one `planner_show` command
i.e. `planner_show year/YEAR && sem/SEMESTER`
returns academic semester(s) having its year and semester matches the given year and semester
- [x] able to include both `year` and `semester` attributes with more complex Boolean expression in one `planner_show` command
i.e. `planner_show year/YEAR && ( sem/SEMESTER || sem/SEMESTER)`
returns academic semester(s) having its year and semester satisfies the Boolean expression
These sub-features will allow the user to only see his/her degree planner(s) that he/she wants. | 1.0 | As a user, I want to see specific degree planner(s). - The user may want to see the degree planners based on certain condition.
Hence, I propose implementing ```planner_show``` command as follows with/without Boolean expression:
- [x] able to show degree planner(s) by year
i.e. `planner_show year/YEAR`
returns academic semester(s) having its year matches the year given
- [x] able to show degree planner(s) by semester
i.e. `planner_show sem/SEMESTER`
returns academic semester(s) having its semester matches the semester given
- [x] able to include both `year` and `semester` attributes in one `planner_show` command
i.e. `planner_show year/YEAR && sem/SEMESTER`
returns academic semester(s) having its year and semester matches the given year and semester
- [x] able to include both `year` and `semester` attributes with more complex Boolean expression in one `planner_show` command
i.e. `planner_show year/YEAR && ( sem/SEMESTER || sem/SEMESTER)`
returns academic semester(s) having its year and semester satisfies the Boolean expression
These sub-features will allow the user to only see his/her degree planner(s) that he/she wants. | non_defect | as a user i want to see specific degree planner s the user may want to see the degree planners based on certain condition hence i propose implementing planner show command as follows with without boolean expression able to show degree planner s by year i e planner show year year returns academic semester s having its year matches the year given able to show degree planner s by semester i e planner show sem semester returns academic semester s having its semester matches the semester given able to include both year and semester attributes in one planner show command i e planner show year year sem semester returns academic semester s having its year and semester matches the given year and semester able to include both year and semester attributes with more complex boolean expression in one planner show command i e planner show year year sem semester sem semester returns academic semester s having its year and semester satisfies the boolean expression these sub features will allow the user to only see his her degree planner s that he she wants | 0 |
147,045 | 11,769,809,902 | IssuesEvent | 2020-03-15 16:27:27 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/feature_controls/ml_security·ts - machine learning feature controls security "before all" hook in "security" | failed-test | A test failed on a tracked branch
```
[Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/dev/shm/workspace/kibana/x-pack/test/functional/apps/machine_learning/feature_controls/ml_security.ts)]
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/3661/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/feature_controls/ml_security·ts","test.name":"machine learning feature controls security \"before all\" hook in \"security\"","test.failCount":1}} --> | 1.0 | Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/feature_controls/ml_security·ts - machine learning feature controls security "before all" hook in "security" - A test failed on a tracked branch
```
[Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/dev/shm/workspace/kibana/x-pack/test/functional/apps/machine_learning/feature_controls/ml_security.ts)]
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/3661/)
<!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/machine_learning/feature_controls/ml_security·ts","test.name":"machine learning feature controls security \"before all\" hook in \"security\"","test.failCount":1}} --> | non_defect | failing test chrome x pack ui functional tests x pack test functional apps machine learning feature controls ml security·ts machine learning feature controls security before all hook in security a test failed on a tracked branch first failure | 0 |
740,857 | 25,771,256,013 | IssuesEvent | 2022-12-09 08:11:06 | internetarchive/openlibrary | https://api.github.com/repos/internetarchive/openlibrary | closed | Covers low on storage: Archive + free 100GB | Type: Bug Priority: 1 Lead: @mekarpeles | We have 68gb of storage left on covers! This will likely last ~2months MAX. Need to compress the images on covers.
### Evidence / Screenshot (if possible)
### Relevant url?
<!-- `https://openlibrary.org/...` -->
### Steps to Reproduce
<!-- What steps caused you to find the bug? -->
1. Go to ...
2. Do ...
<!-- What actually happened after these steps? What did you expect to happen? -->
* Actual:
* Expected:
### Details
- **Logged in (Y/N)?**
- **Browser type/version?**
- **Operating system?**
- **Environment (prod/dev/local)?** prod
<!-- If not sure, put prod -->
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
| 1.0 | Covers low on storage: Archive + free 100GB - We have 68gb of storage left on covers! This will likely last ~2months MAX. Need to compress the images on covers.
### Evidence / Screenshot (if possible)
### Relevant url?
<!-- `https://openlibrary.org/...` -->
### Steps to Reproduce
<!-- What steps caused you to find the bug? -->
1. Go to ...
2. Do ...
<!-- What actually happened after these steps? What did you expect to happen? -->
* Actual:
* Expected:
### Details
- **Logged in (Y/N)?**
- **Browser type/version?**
- **Operating system?**
- **Environment (prod/dev/local)?** prod
<!-- If not sure, put prod -->
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
### Related files
<!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
| non_defect | covers low on storage archive free we have of storage left on covers this will likely last max need to compress the images on covers evidence screenshot if possible relevant url steps to reproduce go to do actual expected details logged in y n browser type version operating system environment prod dev local prod proposal constraints related files stakeholders | 0 |
129,704 | 5,101,742,252 | IssuesEvent | 2017-01-04 16:18:32 | GoogleCloudPlatform/google-cloud-eclipse | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse | closed | Java VM can't start but server stays running | bug low priority | I forced the VM to fail when launching a local server by supplying an undefined argument to the VM. The WTP server doesn't notice this and keeps going.
```
INFO 2016-12-29 12:37:41,007 devappserver2.py:769] Skipping SDK update check.
INFO 2016-12-29 12:37:41,116 api_server.py:205] Starting API server at: http://localhost:52060
INFO 2016-12-29 12:37:41,123 dispatcher.py:197] Starting module "default" running at: http://localhost:8080
INFO 2016-12-29 12:37:41,127 admin_server.py:116] Starting admin server at: http://localhost:8000
ERROR 2016-12-29 12:37:42,270 http_runtime.py:396] bad runtime process port ['']
Unrecognized option: -versub2
Error: Could not create the Java Virtual Machine.
Error: A fatal exception has occurred. Program will exit.
ERROR 2016-12-29 12:37:42,276 http_proxy.py:116] bad runtime process port ['']
INFO 2016-12-29 12:37:42,383 module.py:788] default: "GET /_ah/warmup HTTP/1.1" 500 169
``` | 1.0 | Java VM can't start but server stays running - I forced the VM to fail when launching a local server by supplying an undefined argument to the VM. The WTP server doesn't notice this and keeps going.
```
INFO 2016-12-29 12:37:41,007 devappserver2.py:769] Skipping SDK update check.
INFO 2016-12-29 12:37:41,116 api_server.py:205] Starting API server at: http://localhost:52060
INFO 2016-12-29 12:37:41,123 dispatcher.py:197] Starting module "default" running at: http://localhost:8080
INFO 2016-12-29 12:37:41,127 admin_server.py:116] Starting admin server at: http://localhost:8000
ERROR 2016-12-29 12:37:42,270 http_runtime.py:396] bad runtime process port ['']
Unrecognized option: -versub2
Error: Could not create the Java Virtual Machine.
Error: A fatal exception has occurred. Program will exit.
ERROR 2016-12-29 12:37:42,276 http_proxy.py:116] bad runtime process port ['']
INFO 2016-12-29 12:37:42,383 module.py:788] default: "GET /_ah/warmup HTTP/1.1" 500 169
``` | non_defect | java vm can t start but server stays running i forced the vm to fail when launching a local server by supplying an undefined argument to the vm the wtp server doesn t notice this and keeps going info py skipping sdk update check info api server py starting api server at info dispatcher py starting module default running at info admin server py starting admin server at error http runtime py bad runtime process port unrecognized option error could not create the java virtual machine error a fatal exception has occurred program will exit error http proxy py bad runtime process port info module py default get ah warmup http | 0 |
20,679 | 3,398,841,614 | IssuesEvent | 2015-12-02 07:28:28 | ckaiser79/webtest-rb | https://api.github.com/repos/ckaiser79/webtest-rb | closed | Browser stays open, if configured to autoclose | auto-migrated OpSys-Windows Priority-Low Type-Defect | ```
What steps will reproduce the problem?
1. set autoclose browser to false
2. run sample testcase
What is the expected output? What do you see instead?
- Brwoser stays open, tested with FF
```
Original issue reported on code.google.com by `barny...@gmail.com` on 19 Sep 2014 at 6:06 | 1.0 | Browser stays open, if configured to autoclose - ```
What steps will reproduce the problem?
1. set autoclose browser to false
2. run sample testcase
What is the expected output? What do you see instead?
- Brwoser stays open, tested with FF
```
Original issue reported on code.google.com by `barny...@gmail.com` on 19 Sep 2014 at 6:06 | defect | browser stays open if configured to autoclose what steps will reproduce the problem set autoclose browser to false run sample testcase what is the expected output what do you see instead brwoser stays open tested with ff original issue reported on code google com by barny gmail com on sep at | 1 |
54,347 | 13,615,061,487 | IssuesEvent | 2020-09-23 14:00:27 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | PANIC at zfs_znode.c zfs_znode_sa_init() | Status: Triage Needed Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Linux VM
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04.1 LTS (Focal Fossa)
Linux Kernel | 5.4.0-48-generic
Architecture | x86_64
ZFS Version | 0.8.3-1ubuntu12.4
SPL Version | 0.8.3-1ubuntu12.4
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
A PANIC event is logged in dmesg
### Describe how to reproduce the problem
Unsure
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
```
VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
PANIC at zfs_znode.c:335:zfs_znode_sa_init()
Showing stack for process 300061
CPU: 0 PID: 300061 Comm: BackupPC_nightl Tainted: P D O 5.4.0-48-generic #52-Ubuntu
Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.10.2-1ubuntu1 04/01/2014
Call Trace:
dump_stack+0x6d/0x9a
spl_dumpstack+0x29/0x2b [spl]
spl_panic+0xd4/0x102 [spl]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? dmu_buf_replace_user+0x75/0xa0 [zfs]
? dmu_buf_set_user+0x13/0x20 [zfs]
? dmu_buf_set_user_ie+0x15/0x20 [zfs]
zfs_znode_sa_init.isra.0+0xde/0xf0 [zfs]
zfs_znode_alloc+0x102/0x570 [zfs]
? atomic_cmpxchg+0x16/0x30 [zfs]
? _cond_resched+0x19/0x30
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? aggsum_add+0xca/0xe0 [zfs]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? dbuf_read_impl+0x426/0x610 [zfs]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? atomic64_add+0x12/0x20 [zfs]
? dbuf_read+0x102/0x5f0 [zfs]
? arc_space_consume+0x54/0xe0 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? dnode_rele_and_unlock+0x68/0xc0 [zfs]
? atomic_cmpxchg+0x16/0x30 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? RW_WRITE_HELD+0xd/0x30 [zfs]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? atomic_dec_return+0x9/0x10 [zfs]
zfs_zget+0x24a/0x290 [zfs]
zfs_dirent_lock+0x41a/0x5a0 [zfs]
zfs_dirlook+0x90/0x2b0 [zfs]
zfs_lookup+0x202/0x3b0 [zfs]
zpl_lookup+0x94/0x210 [zfs]
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
__lookup_slow+0x92/0x160
lookup_slow+0x3b/0x60
walk_component+0x1da/0x360
? link_path_walk.part.0+0x6d/0x550
path_lookupat.isra.0+0x80/0x230
? kmem_cache_free+0x288/0x2b0
filename_lookup+0xae/0x170
? __check_object_size+0x13f/0x150
? strncpy_from_user+0x4c/0x150
user_path_at_empty+0x3a/0x50
vfs_statx+0x7d/0xe0
__do_sys_newlstat+0x3e/0x80
__x64_sys_newlstat+0x16/0x20
do_syscall_64+0x57/0x190
entry_SYSCALL_64_after_hwframe+0x44/0xa9
RIP: 0033:0x7fef462e755a
Code: ff ff ff c3 66 2e 0f 1f 84 00 00 00 00 00 66 90 f3 0f 1e fa 41 89 f8 48 89 f7 48 89 d6 41 83 f8 01 77 2d b8 06 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 06 c3 0f 1f 44 00 00 48 8b 15 01 a9 0d 00 f7
RSP: 002b:00007ffc2c409ad8 EFLAGS: 00000246 ORIG_RAX: 0000000000000006
RAX: ffffffffffffffda RBX: 0000561fac283270 RCX: 00007fef462e755a
RDX: 0000561fab6d84b8 RSI: 0000561fab6d84b8 RDI: 0000561fabe924f0
RBP: 0000561fab6d82a0 R08: 0000000000000001 R09: aaaaaaaaaaaaaaab
R10: 0000561fac2df8d0 R11: 0000000000000246 R12: 0000561fac2df8d8
R13: 0000561fabe924f0 R14: 0000561fa980559e R15: 0000000000000000
```
| 1.0 | PANIC at zfs_znode.c zfs_znode_sa_init() - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Linux VM
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04.1 LTS (Focal Fossa)
Linux Kernel | 5.4.0-48-generic
Architecture | x86_64
ZFS Version | 0.8.3-1ubuntu12.4
SPL Version | 0.8.3-1ubuntu12.4
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
A PANIC event is logged in dmesg
### Describe how to reproduce the problem
Unsure
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
```
VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
PANIC at zfs_znode.c:335:zfs_znode_sa_init()
Showing stack for process 300061
CPU: 0 PID: 300061 Comm: BackupPC_nightl Tainted: P D O 5.4.0-48-generic #52-Ubuntu
Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.10.2-1ubuntu1 04/01/2014
Call Trace:
dump_stack+0x6d/0x9a
spl_dumpstack+0x29/0x2b [spl]
spl_panic+0xd4/0x102 [spl]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? dmu_buf_replace_user+0x75/0xa0 [zfs]
? dmu_buf_set_user+0x13/0x20 [zfs]
? dmu_buf_set_user_ie+0x15/0x20 [zfs]
zfs_znode_sa_init.isra.0+0xde/0xf0 [zfs]
zfs_znode_alloc+0x102/0x570 [zfs]
? atomic_cmpxchg+0x16/0x30 [zfs]
? _cond_resched+0x19/0x30
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? aggsum_add+0xca/0xe0 [zfs]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? dbuf_read_impl+0x426/0x610 [zfs]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? atomic64_add+0x12/0x20 [zfs]
? dbuf_read+0x102/0x5f0 [zfs]
? arc_space_consume+0x54/0xe0 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? __raw_spin_unlock+0x9/0x10 [zfs]
? dnode_rele_and_unlock+0x68/0xc0 [zfs]
? atomic_cmpxchg+0x16/0x30 [zfs]
? do_raw_spin_unlock+0x9/0x10 [zfs]
? RW_WRITE_HELD+0xd/0x30 [zfs]
? atomic_sub_return.constprop.0+0xd/0x20 [zfs]
? atomic_dec_return+0x9/0x10 [zfs]
zfs_zget+0x24a/0x290 [zfs]
zfs_dirent_lock+0x41a/0x5a0 [zfs]
zfs_dirlook+0x90/0x2b0 [zfs]
zfs_lookup+0x202/0x3b0 [zfs]
zpl_lookup+0x94/0x210 [zfs]
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
__lookup_slow+0x92/0x160
lookup_slow+0x3b/0x60
walk_component+0x1da/0x360
? link_path_walk.part.0+0x6d/0x550
path_lookupat.isra.0+0x80/0x230
? kmem_cache_free+0x288/0x2b0
filename_lookup+0xae/0x170
? __check_object_size+0x13f/0x150
? strncpy_from_user+0x4c/0x150
user_path_at_empty+0x3a/0x50
vfs_statx+0x7d/0xe0
__do_sys_newlstat+0x3e/0x80
__x64_sys_newlstat+0x16/0x20
do_syscall_64+0x57/0x190
entry_SYSCALL_64_after_hwframe+0x44/0xa9
RIP: 0033:0x7fef462e755a
Code: ff ff ff c3 66 2e 0f 1f 84 00 00 00 00 00 66 90 f3 0f 1e fa 41 89 f8 48 89 f7 48 89 d6 41 83 f8 01 77 2d b8 06 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 06 c3 0f 1f 44 00 00 48 8b 15 01 a9 0d 00 f7
RSP: 002b:00007ffc2c409ad8 EFLAGS: 00000246 ORIG_RAX: 0000000000000006
RAX: ffffffffffffffda RBX: 0000561fac283270 RCX: 00007fef462e755a
RDX: 0000561fab6d84b8 RSI: 0000561fab6d84b8 RDI: 0000561fabe924f0
RBP: 0000561fab6d82a0 R08: 0000000000000001 R09: aaaaaaaaaaaaaaab
R10: 0000561fac2df8d0 R11: 0000000000000246 R12: 0000561fac2df8d8
R13: 0000561fabe924f0 R14: 0000561fa980559e R15: 0000000000000000
```
| defect | panic at zfs znode c zfs znode sa init thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type linux vm distribution name ubuntu distribution version lts focal fossa linux kernel generic architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing a panic event is logged in dmesg describe how to reproduce the problem unsure include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with verify sa handle get from db zfsvfs z os db zp sa hdl shared zp z sa hdl failed panic at zfs znode c zfs znode sa init showing stack for process cpu pid comm backuppc nightl tainted p d o generic ubuntu hardware name qemu standard pc piix bios call trace dump stack spl dumpstack spl panic atomic sub return constprop do raw spin unlock raw spin unlock dmu buf replace user dmu buf set user dmu buf set user ie zfs znode sa init isra zfs znode alloc atomic cmpxchg cond resched do raw spin unlock raw spin unlock aggsum add atomic sub return constprop do raw spin unlock raw spin unlock dbuf read impl atomic sub return constprop add dbuf read arc space consume do raw spin unlock do raw spin unlock raw spin unlock dnode rele and unlock atomic cmpxchg do raw spin unlock rw write held atomic sub return constprop atomic dec return zfs zget zfs dirent lock zfs dirlook zfs lookup zpl lookup switch to asm switch to asm switch to asm switch to asm lookup slow lookup slow walk component link path walk part path lookupat isra kmem cache free filename lookup check object size strncpy from user user path at empty vfs statx do sys newlstat sys newlstat do syscall entry syscall after hwframe rip code ff ff ff fa ff ff rsp eflags orig rax rax ffffffffffffffda rbx rcx rdx rsi rdi rbp aaaaaaaaaaaaaaab | 1 |
18,172 | 3,031,498,503 | IssuesEvent | 2015-08-04 23:24:28 | googlei18n/noto-fonts | https://api.github.com/repos/googlei18n/noto-fonts | opened | Inscriptional Parthian: undesired glyph intersections | Script-Parthian Type-Defect | Imported from Google Code issue #16 created by roozbeh@google.com on 2013-11-14T03:02:56.000Z:
The sequence <Nun, Shin> results in the tail of the Nun intersecting with the Shin.
The sequence <Nun, Sadhe> results in the tails of the two letters intersecting.
The sequence <Sadhe, Sadhe> results in the tails of the two letters intersecting.
The sequence <Sadhe, Nun, Daleth> results in the tails of Sadhe and Nun intersecting.
The sequence <Sadhe, Nun, Sadhe> results in the tails of all three letters intersecting. | 1.0 | Inscriptional Parthian: undesired glyph intersections - Imported from Google Code issue #16 created by roozbeh@google.com on 2013-11-14T03:02:56.000Z:
The sequence <Nun, Shin> results in the tail of the Nun intersecting with the Shin.
The sequence <Nun, Sadhe> results in the tails of the two letters intersecting.
The sequence <Sadhe, Sadhe> results in the tails of the two letters intersecting.
The sequence <Sadhe, Nun, Daleth> results in the tails of Sadhe and Nun intersecting.
The sequence <Sadhe, Nun, Sadhe> results in the tails of all three letters intersecting. | defect | inscriptional parthian undesired glyph intersections imported from google code issue created by roozbeh google com on the sequence results in the tail of the nun intersecting with the shin the sequence results in the tails of the two letters intersecting the sequence results in the tails of the two letters intersecting the sequence results in the tails of sadhe and nun intersecting the sequence results in the tails of all three letters intersecting | 1 |
231,822 | 25,551,989,779 | IssuesEvent | 2022-11-30 01:05:04 | RG4421/atlasdb | https://api.github.com/repos/RG4421/atlasdb | opened | CVE-2022-41854 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2022-41854 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.23.jar</b>, <b>snakeyaml-1.26.jar</b>, <b>snakeyaml-1.24.jar</b></p></summary>
<p>
<details><summary><b>snakeyaml-1.23.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /lock-impl/build.gradle</p>
<p>Path to vulnerable library: /20210226193332_TENYLC/downloadResource_RSJUCV/20210226194911/snakeyaml-1.23.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.23.jar** (Vulnerable Library)
</details>
<details><summary><b>snakeyaml-1.26.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /atlasdb-service/build.gradle</p>
<p>Path to vulnerable library: /timelock-api/build/conjureJava/lib/snakeyaml-1.26.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar,/canner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.26.jar** (Vulnerable Library)
</details>
<details><summary><b>snakeyaml-1.24.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to vulnerable library: /timelock-api/build/conjureCompiler/lib/snakeyaml-1.24.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.24.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Snakeyaml to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-11-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41854>CVE-2022-41854</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| True | CVE-2022-41854 (Medium) detected in multiple libraries - ## CVE-2022-41854 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.23.jar</b>, <b>snakeyaml-1.26.jar</b>, <b>snakeyaml-1.24.jar</b></p></summary>
<p>
<details><summary><b>snakeyaml-1.23.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /lock-impl/build.gradle</p>
<p>Path to vulnerable library: /20210226193332_TENYLC/downloadResource_RSJUCV/20210226194911/snakeyaml-1.23.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.23.jar** (Vulnerable Library)
</details>
<details><summary><b>snakeyaml-1.26.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /atlasdb-service/build.gradle</p>
<p>Path to vulnerable library: /timelock-api/build/conjureJava/lib/snakeyaml-1.26.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar,/canner/.gradle/caches/modules-2/files-2.1/org.yaml/snakeyaml/1.26/a78a8747147d2c5807683e76ec2b633e95c14fe9/snakeyaml-1.26.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.26.jar** (Vulnerable Library)
</details>
<details><summary><b>snakeyaml-1.24.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to vulnerable library: /timelock-api/build/conjureCompiler/lib/snakeyaml-1.24.jar</p>
<p>
Dependency Hierarchy:
- :x: **snakeyaml-1.24.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using Snakeyaml to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack.
<p>Publish Date: 2022-11-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41854>CVE-2022-41854</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| non_defect | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries snakeyaml jar snakeyaml jar snakeyaml jar snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file lock impl build gradle path to vulnerable library tenylc downloadresource rsjucv snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file atlasdb service build gradle path to vulnerable library timelock api build conjurejava lib snakeyaml jar home wss scanner gradle caches modules files org yaml snakeyaml snakeyaml jar canner gradle caches modules files org yaml snakeyaml snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to vulnerable library timelock api build conjurecompiler lib snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library found in base branch develop vulnerability details those using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stack overflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href | 0 |
4,424 | 2,610,093,874 | IssuesEvent | 2015-02-26 18:28:20 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳韩方科颜怎么样 | auto-migrated Priority-Medium Type-Defect | ```
深圳韩方科颜怎么样【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:11 | 1.0 | 深圳韩方科颜怎么样 - ```
深圳韩方科颜怎么样【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:11 | defect | 深圳韩方科颜怎么样 深圳韩方科颜怎么样【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国�� �方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩� ��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹” 健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专�� �治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的� ��痘。 original issue reported on code google com by szft com on may at | 1 |
13,093 | 2,732,897,000 | IssuesEvent | 2015-04-17 10:04:07 | tiku01/oryx-editor | https://api.github.com/repos/tiku01/oryx-editor | closed | Repo2: Make tableview sortable | auto-migrated Priority-Low Type-Defect | ```
What steps will reproduce the problem?
1. Go in Repo to Tableview
2. Klick on a column header
3.
What is the expected output?
- Sorted (like in excel) by the field type selected.
What do you see instead?
- nothing happens
Please provide any additional information below.
- Important to sort for model type, creation date or owner
```
Original issue reported on code.google.com by `alexande...@googlemail.com` on 23 Nov 2008 at 12:37 | 1.0 | Repo2: Make tableview sortable - ```
What steps will reproduce the problem?
1. Go in Repo to Tableview
2. Klick on a column header
3.
What is the expected output?
- Sorted (like in excel) by the field type selected.
What do you see instead?
- nothing happens
Please provide any additional information below.
- Important to sort for model type, creation date or owner
```
Original issue reported on code.google.com by `alexande...@googlemail.com` on 23 Nov 2008 at 12:37 | defect | make tableview sortable what steps will reproduce the problem go in repo to tableview klick on a column header what is the expected output sorted like in excel by the field type selected what do you see instead nothing happens please provide any additional information below important to sort for model type creation date or owner original issue reported on code google com by alexande googlemail com on nov at | 1 |
143,252 | 19,177,902,236 | IssuesEvent | 2021-12-04 00:03:14 | samq-ghdemo/js-monorepo | https://api.github.com/repos/samq-ghdemo/js-monorepo | opened | CVE-2020-35149 (Medium) detected in mquery-1.6.3.tgz | security vulnerability | ## CVE-2020-35149 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mquery-1.6.3.tgz</b></p></summary>
<p>Expressive query building for MongoDB</p>
<p>Library home page: <a href="https://registry.npmjs.org/mquery/-/mquery-1.6.3.tgz">https://registry.npmjs.org/mquery/-/mquery-1.6.3.tgz</a></p>
<p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p>
<p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/mquery/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-4.2.4.tgz (Root Library)
- :x: **mquery-1.6.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lib/utils.js in mquery before 3.2.3 allows a pollution attack because a special property (e.g., __proto__) can be copied during a merge or clone operation.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35149>CVE-2020-35149</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/aheckmann/mquery/releases/tag/3.2.3">https://github.com/aheckmann/mquery/releases/tag/3.2.3</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: 3.2.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mquery","packageVersion":"1.6.3","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"mongoose:4.2.4;mquery:1.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.2.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-35149","vulnerabilityDetails":"lib/utils.js in mquery before 3.2.3 allows a pollution attack because a special property (e.g., __proto__) can be copied during a merge or clone operation.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35149","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-35149 (Medium) detected in mquery-1.6.3.tgz - ## CVE-2020-35149 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mquery-1.6.3.tgz</b></p></summary>
<p>Expressive query building for MongoDB</p>
<p>Library home page: <a href="https://registry.npmjs.org/mquery/-/mquery-1.6.3.tgz">https://registry.npmjs.org/mquery/-/mquery-1.6.3.tgz</a></p>
<p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p>
<p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/mquery/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-4.2.4.tgz (Root Library)
- :x: **mquery-1.6.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lib/utils.js in mquery before 3.2.3 allows a pollution attack because a special property (e.g., __proto__) can be copied during a merge or clone operation.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35149>CVE-2020-35149</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/aheckmann/mquery/releases/tag/3.2.3">https://github.com/aheckmann/mquery/releases/tag/3.2.3</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: 3.2.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mquery","packageVersion":"1.6.3","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"mongoose:4.2.4;mquery:1.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.2.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-35149","vulnerabilityDetails":"lib/utils.js in mquery before 3.2.3 allows a pollution attack because a special property (e.g., __proto__) can be copied during a merge or clone operation.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35149","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_defect | cve medium detected in mquery tgz cve medium severity vulnerability vulnerable library mquery tgz expressive query building for mongodb library home page a href path to dependency file js monorepo nodejs goof package json path to vulnerable library js monorepo nodejs goof node modules mquery package json dependency hierarchy mongoose tgz root library x mquery tgz vulnerable library found in head commit a href found in base branch main vulnerability details lib utils js in mquery before allows a pollution attack because a special property e g proto can be copied during a merge or clone operation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree mongoose mquery isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails lib utils js in mquery before allows a pollution attack because a special property e g proto can be copied during a merge or clone operation vulnerabilityurl | 0 |
637,892 | 20,681,208,422 | IssuesEvent | 2022-03-10 14:04:52 | if-mentor/nuxt3_todo | https://api.github.com/repos/if-mentor/nuxt3_todo | closed | [ UI ] Todo編集画面の作成 | priority high | ## 実装内容の説明
- 編集ページの作成。
- TODOを入力するフォームを用意してください。
## 実装予定のページ・コンポーネントのデザイン画像もしくはURL

https://www.figma.com/file/3Db2buFg9YrFTaEJy8r3Cz/Todo%E3%82%A2%E3%83%97%E3%83%AA%E9%96%8B%E7%99%BA?node-id=48%3A5
## TASK
- [ ] ページの作成
- [ ] フォームの追加
- [ ] ボタンの追加
## その他・備考
-
| 1.0 | [ UI ] Todo編集画面の作成 - ## 実装内容の説明
- 編集ページの作成。
- TODOを入力するフォームを用意してください。
## 実装予定のページ・コンポーネントのデザイン画像もしくはURL

https://www.figma.com/file/3Db2buFg9YrFTaEJy8r3Cz/Todo%E3%82%A2%E3%83%97%E3%83%AA%E9%96%8B%E7%99%BA?node-id=48%3A5
## TASK
- [ ] ページの作成
- [ ] フォームの追加
- [ ] ボタンの追加
## その他・備考
-
| non_defect | todo編集画面の作成 実装内容の説明 編集ページの作成。 todoを入力するフォームを用意してください。 実装予定のページ・コンポーネントのデザイン画像もしくはurl task ページの作成 フォームの追加 ボタンの追加 その他・備考 | 0 |
50,695 | 13,187,683,785 | IssuesEvent | 2020-08-13 04:13:32 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | [wavedeform] example is broken (Trac #1195) | Migrated from Trac combo reconstruction defect | It tries to run WaveCalibrator, but the syntax is wrong. Fixing that produces other errors, so I think something changed within the last few years.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1195">https://code.icecube.wisc.edu/ticket/1195</a>, reported by david.schultz and owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "It tries to run WaveCalibrator, but the syntax is wrong. Fixing that produces other errors, so I think something changed within the last few years.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[wavedeform] example is broken",
"priority": "blocker",
"keywords": "",
"time": "2015-08-19T17:11:14",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [wavedeform] example is broken (Trac #1195) - It tries to run WaveCalibrator, but the syntax is wrong. Fixing that produces other errors, so I think something changed within the last few years.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1195">https://code.icecube.wisc.edu/ticket/1195</a>, reported by david.schultz and owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "It tries to run WaveCalibrator, but the syntax is wrong. Fixing that produces other errors, so I think something changed within the last few years.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[wavedeform] example is broken",
"priority": "blocker",
"keywords": "",
"time": "2015-08-19T17:11:14",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| defect | example is broken trac it tries to run wavecalibrator but the syntax is wrong fixing that produces other errors so i think something changed within the last few years migrated from json status closed changetime description it tries to run wavecalibrator but the syntax is wrong fixing that produces other errors so i think something changed within the last few years reporter david schultz cc resolution fixed ts component combo reconstruction summary example is broken priority blocker keywords time milestone owner jbraun type defect | 1 |
49,504 | 13,187,222,279 | IssuesEvent | 2020-08-13 02:44:10 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | Summary of remaining static analysis Issues (Trac #1549) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1549">https://code.icecube.wisc.edu/ticket/1549</a>, reported by kjmeagher and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2017-04-24T08:39:36",
"description": "\n== Code that is clearly producing incorrect results ==\n`CascadeVariables` returns uninitilized data see bug #1537\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-7461e5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-bb648a.html#EndPath\n\n== Code that need improvement ==\n\nCode assumes that the file being read starts with a `#` and contains the number of strings \nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-6b756c.html#EndPath\n\ncode assumes 8 parameters when the number of parameters is a variable\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-1e32f9.html#EndPath\n\ncode path is confusing and has lots of `#ifdefs`\ntoprec/private/toprec/laputop/I3LaputopParametrization.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-1a61fb.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-cfabfc.html#EndPath\n\n== Code I gave up trying to understand the problem==\nPROPOSAL/private/PROPOSAL/Amanda.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-19a21f.html#EndPath\n\nVHESelfVeto/private/clipper/clipper.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-47fef5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-da4893.html#EndPath\n\nfill-ratio/private/fill-ratio/I3FillRatioLite.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-146975.html#EndPath\n\nicetray/private/icetray/I3Tray.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-fe2e6e.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-873438.html#EndPath\n\nipdf/private/pybindings/Likelihood.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-6148e1.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-7d3a26.html#EndPath\n\nmue/private/mue/llhreco.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-b0852c.html#EndPath\n\ndead assignment is the result of `decode()` but i can't tell if `decode()` has side effects or not\npayload-parsing/private/payload-parsing/dump-raw-deltacompressed.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-c9677b.html#EndPath\n\n\n\ntpx/private/tpx/I3IceTopBaselineModule.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-aaed35.html#EndPath\n\n\n== copy and paste files with too many problems to list ==\n\ng4-tankresponse/private/g4-tankresponse/triangle/triangle.c\nlilliput/private/minimizer/minuit/TMinuit1.cxx\n\n== Memory leaks ==\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopDetectorConstruction.cxx\t\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-5704c1.html#EndPath\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopTank.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-271865.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-4e6179.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-e17000.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-e87022.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-f536c8.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-f71588.html#EndPath\n\nmue/private/mue/I3mue.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-d05444.html#EndPath\n\n== False Positives ==\ncomplains about an impossible code branch\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-141e17.html#EndPath\n\nclang confuses unix's `open` with `I3::dataio::open` see #1544\nclsim/private/make_safeprimes/convert.cxx\ndataio/private/dataio/I3MultiWriter.cxx\ndataio/private/dataio/I3Writer.cxx\ndataio/private/pybindings/I3File.cxx\ndataio/private/shovel/Model.cxx\ndataio/private/shovel/Model.cxx\ndataio/private/shovel/Model.cxx\nsteamshovel/private/steamshovel/FileService.cpp\n\nclang dosn't understand `G4Exception()`\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-1d0a97.html#EndPath\n\nclang doesn't understand Qt memory management see #1545\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-4ca3c9.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-5c5809.html#EndPath",
"reporter": "kjmeagher",
"cc": "david.schultz, nega, olivas",
"resolution": "fixed",
"_ts": "1493023176837564",
"component": "cmake",
"summary": "Summary of remaining static analysis Issues",
"priority": "normal",
"keywords": "",
"time": "2016-02-15T10:27:59",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Summary of remaining static analysis Issues (Trac #1549) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1549">https://code.icecube.wisc.edu/ticket/1549</a>, reported by kjmeagher and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2017-04-24T08:39:36",
"description": "\n== Code that is clearly producing incorrect results ==\n`CascadeVariables` returns uninitilized data see bug #1537\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-7461e5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-bb648a.html#EndPath\n\n== Code that need improvement ==\n\nCode assumes that the file being read starts with a `#` and contains the number of strings \nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-6b756c.html#EndPath\n\ncode assumes 8 parameters when the number of parameters is a variable\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-1e32f9.html#EndPath\n\ncode path is confusing and has lots of `#ifdefs`\ntoprec/private/toprec/laputop/I3LaputopParametrization.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-1a61fb.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-cfabfc.html#EndPath\n\n== Code I gave up trying to understand the problem==\nPROPOSAL/private/PROPOSAL/Amanda.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-19a21f.html#EndPath\n\nVHESelfVeto/private/clipper/clipper.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-47fef5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-da4893.html#EndPath\n\nfill-ratio/private/fill-ratio/I3FillRatioLite.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-146975.html#EndPath\n\nicetray/private/icetray/I3Tray.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-fe2e6e.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-873438.html#EndPath\n\nipdf/private/pybindings/Likelihood.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-6148e1.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-7d3a26.html#EndPath\n\nmue/private/mue/llhreco.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-b0852c.html#EndPath\n\ndead assignment is the result of `decode()` but i can't tell if `decode()` has side effects or not\npayload-parsing/private/payload-parsing/dump-raw-deltacompressed.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-c9677b.html#EndPath\n\n\n\ntpx/private/tpx/I3IceTopBaselineModule.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-aaed35.html#EndPath\n\n\n== copy and paste files with too many problems to list ==\n\ng4-tankresponse/private/g4-tankresponse/triangle/triangle.c\nlilliput/private/minimizer/minuit/TMinuit1.cxx\n\n== Memory leaks ==\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopDetectorConstruction.cxx\t\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-5704c1.html#EndPath\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopTank.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-271865.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-4e6179.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-e17000.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-e87022.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-f536c8.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-f71588.html#EndPath\n\nmue/private/mue/I3mue.cxx\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-d05444.html#EndPath\n\n== False Positives ==\ncomplains about an impossible code branch\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-141e17.html#EndPath\n\nclang confuses unix's `open` with `I3::dataio::open` see #1544\nclsim/private/make_safeprimes/convert.cxx\ndataio/private/dataio/I3MultiWriter.cxx\ndataio/private/dataio/I3Writer.cxx\ndataio/private/pybindings/I3File.cxx\ndataio/private/shovel/Model.cxx\ndataio/private/shovel/Model.cxx\ndataio/private/shovel/Model.cxx\nsteamshovel/private/steamshovel/FileService.cpp\n\nclang dosn't understand `G4Exception()`\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-1d0a97.html#EndPath\n\nclang doesn't understand Qt memory management see #1545\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-4ca3c9.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-02-14-030225-7689-1/report-5c5809.html#EndPath",
"reporter": "kjmeagher",
"cc": "david.schultz, nega, olivas",
"resolution": "fixed",
"_ts": "1493023176837564",
"component": "cmake",
"summary": "Summary of remaining static analysis Issues",
"priority": "normal",
"keywords": "",
"time": "2016-02-15T10:27:59",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | summary of remaining static analysis issues trac migrated from json status closed changetime description n code that is clearly producing incorrect results n cascadevariables returns uninitilized data see bug n code that need improvement n ncode assumes that the file being read starts with a and contains the number of strings n assumes parameters when the number of parameters is a variable n path is confusing and has lots of ifdefs ntoprec private toprec laputop cxx n code i gave up trying to understand the problem nproposal private proposal amanda cxx n assignment is the result of decode but i can t tell if decode has side effects or not npayload parsing private payload parsing dump raw deltacompressed cxx n copy and paste files with too many problems to list n tankresponse private tankresponse triangle triangle c nlilliput private minimizer minuit cxx n n memory leaks tankresponse private tankresponse cxx t n false positives ncomplains about an impossible code branch n confuses unix s open with dataio open see nclsim private make safeprimes convert cxx ndataio private dataio cxx ndataio private dataio cxx ndataio private pybindings cxx ndataio private shovel model cxx ndataio private shovel model cxx ndataio private shovel model cxx nsteamshovel private steamshovel fileservice cpp n nclang dosn t understand n doesn t understand qt memory management see n reporter kjmeagher cc david schultz nega olivas resolution fixed ts component cmake summary summary of remaining static analysis issues priority normal keywords time milestone long term future owner nega type defect | 1 |
28,860 | 23,532,671,307 | IssuesEvent | 2022-08-19 16:56:05 | lampepfl/dotty | https://api.github.com/repos/lampepfl/dotty | opened | CI test run fails but CI passes and missed a compiler regression | itype:bug area:infrastructure regression | The following issue code already exists in the test CI, but it does not cause the CI to fail.
For example:
```
2022-07-20T11:31:07.9036071Z [info] Test dotty.tools.dotc.FromTastyTests.runTestFromTasty started
2022-07-20T11:33:09.2040111Z No tests available under target - erroneous test?
2022-07-20T11:33:09.8939791Z [info] Test dotty.tools.dotc.FromTastyTests.posTestFromTasty started
2022-07-20T11:35:11.2843408Z No tests available under target - erroneous test?
```
As a result, we missed a regression.
See [`tests/run/backwardsCompat-implicitParens/A_1_c3.0.2.scala`
](https://github.com/lampepfl/dotty/blob/main/tests/run/backwardsCompat-implicitParens/A_1_c3.0.2.scala)
Related to the work done in #14840 and followed by #15679 (the last probably introduced the regression)
## Compiler version
v3.2.0-RC4
## Minimized code
```Scala
class Bar(using x: Int)(y: String):
override def toString = "Bar"
object Bar:
given Int = 1
inline def foo =
println(new Bar()(""))
println(Bar()(""))
```
## Output
```scala
missing argument for parameter y of constructor Bar in class Bar: (y: String): Bar
```
## Expectation
No error, and the CI should be fixed that this error reflects in the underlying CI fail/success indication. | 1.0 | CI test run fails but CI passes and missed a compiler regression - The following issue code already exists in the test CI, but it does not cause the CI to fail.
For example:
```
2022-07-20T11:31:07.9036071Z [info] Test dotty.tools.dotc.FromTastyTests.runTestFromTasty started
2022-07-20T11:33:09.2040111Z No tests available under target - erroneous test?
2022-07-20T11:33:09.8939791Z [info] Test dotty.tools.dotc.FromTastyTests.posTestFromTasty started
2022-07-20T11:35:11.2843408Z No tests available under target - erroneous test?
```
As a result, we missed a regression.
See [`tests/run/backwardsCompat-implicitParens/A_1_c3.0.2.scala`
](https://github.com/lampepfl/dotty/blob/main/tests/run/backwardsCompat-implicitParens/A_1_c3.0.2.scala)
Related to the work done in #14840 and followed by #15679 (the last probably introduced the regression)
## Compiler version
v3.2.0-RC4
## Minimized code
```Scala
class Bar(using x: Int)(y: String):
override def toString = "Bar"
object Bar:
given Int = 1
inline def foo =
println(new Bar()(""))
println(Bar()(""))
```
## Output
```scala
missing argument for parameter y of constructor Bar in class Bar: (y: String): Bar
```
## Expectation
No error, and the CI should be fixed that this error reflects in the underlying CI fail/success indication. | non_defect | ci test run fails but ci passes and missed a compiler regression the following issue code already exists in the test ci but it does not cause the ci to fail for example test dotty tools dotc fromtastytests runtestfromtasty started no tests available under target erroneous test test dotty tools dotc fromtastytests postestfromtasty started no tests available under target erroneous test as a result we missed a regression see tests run backwardscompat implicitparens a scala related to the work done in and followed by the last probably introduced the regression compiler version minimized code scala class bar using x int y string override def tostring bar object bar given int inline def foo println new bar println bar output scala missing argument for parameter y of constructor bar in class bar y string bar expectation no error and the ci should be fixed that this error reflects in the underlying ci fail success indication | 0 |
12,086 | 2,684,686,938 | IssuesEvent | 2015-03-29 08:09:44 | netty/netty | https://api.github.com/repos/netty/netty | closed | Cannot subclass HttpHeaders and build with JDK 7 with maven build of 4.0.26.Final | defect | It seems the build of Netty 4.0.26.Final from Maven is built with JDK 8, which adds default methods to collections classes that reference classes that don't exist in JDK 7. That's binary compatible, but not source compatible, as follows:
I am stuck with JDK 7 for a while longer (thanks @schue ) for Actuer. As of Netty 4.0.26.Final my [build is broken](http://timboudreau.com/builds/job/mastfrog-parent/703/console) because javac sees `spliterator` methods and classes in `HttpHeaders`. To fix it I'm staying with Netty 4.0.25 for now for Acteur, but that's not a long-term solution.
Since AFAIK, Netty 4 is buildable on JDK 7, is it possible for JARs released to Maven Central also to be built with JDK 7?
I imagine I am not the only one in this situation. | 1.0 | Cannot subclass HttpHeaders and build with JDK 7 with maven build of 4.0.26.Final - It seems the build of Netty 4.0.26.Final from Maven is built with JDK 8, which adds default methods to collections classes that reference classes that don't exist in JDK 7. That's binary compatible, but not source compatible, as follows:
I am stuck with JDK 7 for a while longer (thanks @schue ) for Actuer. As of Netty 4.0.26.Final my [build is broken](http://timboudreau.com/builds/job/mastfrog-parent/703/console) because javac sees `spliterator` methods and classes in `HttpHeaders`. To fix it I'm staying with Netty 4.0.25 for now for Acteur, but that's not a long-term solution.
Since AFAIK, Netty 4 is buildable on JDK 7, is it possible for JARs released to Maven Central also to be built with JDK 7?
I imagine I am not the only one in this situation. | defect | cannot subclass httpheaders and build with jdk with maven build of final it seems the build of netty final from maven is built with jdk which adds default methods to collections classes that reference classes that don t exist in jdk that s binary compatible but not source compatible as follows i am stuck with jdk for a while longer thanks schue for actuer as of netty final my because javac sees spliterator methods and classes in httpheaders to fix it i m staying with netty for now for acteur but that s not a long term solution since afaik netty is buildable on jdk is it possible for jars released to maven central also to be built with jdk i imagine i am not the only one in this situation | 1 |
474,853 | 13,683,409,103 | IssuesEvent | 2020-09-30 01:45:21 | dojot/dojot | https://api.github.com/repos/dojot/dojot | closed | [GUI] Incorrect date in flows cards | Priority:Medium Team:Frontend Type:Bug :bug: | Timestamp conversion is incorrect in flows cards.
- flows list


- device list


- template list

Affected version: **63.1-20190705** | 1.0 | [GUI] Incorrect date in flows cards - Timestamp conversion is incorrect in flows cards.
- flows list


- device list


- template list

Affected version: **63.1-20190705** | non_defect | incorrect date in flows cards timestamp conversion is incorrect in flows cards flows list device list template list affected version | 0 |
198,511 | 6,973,783,123 | IssuesEvent | 2017-12-11 21:45:45 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.bet365.com - site is not usable | browser-firefox priority-normal type-stylo | <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @extra_label: type-stylo -->
**URL**: https://www.bet365.com/?affiliate=365_229589&cb=105812013192#/AS/B1/
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Site is not usable
**Description**: k
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/12/88428529-e33a-444e-b984-36db95505d92.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.bet365.com - site is not usable - <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @extra_label: type-stylo -->
**URL**: https://www.bet365.com/?affiliate=365_229589&cb=105812013192#/AS/B1/
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Site is not usable
**Description**: k
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/12/88428529-e33a-444e-b984-36db95505d92.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_defect | site is not usable url browser version firefox operating system windows tested another browser unknown problem type site is not usable description k steps to reproduce layout css servo enabled true from with ❤️ | 0 |
232,435 | 25,578,651,440 | IssuesEvent | 2022-12-01 01:15:38 | artkamote/examples | https://api.github.com/repos/artkamote/examples | opened | CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz | security vulnerability | ## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /google-node-typescript-http-endpoint/package.json</p>
<p>Path to vulnerable library: /google-node-typescript-http-endpoint/node_modules/decode-uri-component/package.json,/aws-node-signed-uploads/node_modules/decode-uri-component/package.json,/azure-node-telegram-bot/node_modules/decode-uri-component/package.json,/aws-node-typescript-nest/node_modules/decode-uri-component/package.json,/aws-python-auth0-custom-authorizers-api/node_modules/decode-uri-component/package.json,/aws-rust-simple-http-endpoint/node_modules/decode-uri-component/package.json,/azure-node-typescript-servicebus-trigger-endpoint/node_modules/decode-uri-component/package.json,/aws-node-typescript-apollo-lambda/node_modules/decode-uri-component/package.json,/aws-node-fullstack/frontend/node_modules/decode-uri-component/package.json,/aws-node-dynamic-image-resizer/node_modules/decode-uri-component/package.json,/aws-node-rest-api-typescript-simple/node_modules/decode-uri-component/package.json,/azure-node-line-bot/node_modules/decode-uri-component/package.json,/aws-node-function-compiled-with-babel/node_modules/decode-uri-component/package.json,/aws-node-github-check/node_modules/decode-uri-component/package.json,/aws-node-typescript-sqs-standard/node_modules/decode-uri-component/package.json,/aws-node-vue-nuxt-ssr/node_modules/decode-uri-component/package.json,/aws-node-http-api-typescript/node_modules/decode-uri-component/package.json,/aws-node-typescript-kinesis/node_modules/decode-uri-component/package.json,/aws-golang-dynamo-stream-to-elasticsearch/node_modules/decode-uri-component/package.json,/aws-golang-auth-examples/node_modules/decode-uri-component/package.json,/aws-node-oauth-dropbox-api/node_modules/decode-uri-component/package.json,/aws-node-rest-api-typescript/node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- webpack-4.28.3.tgz
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- source-map-resolve-0.5.2.tgz
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/artkamote/examples/commit/dcbe4aefe4b3685f4b15493a01db0f19b118a0c4">dcbe4aefe4b3685f4b15493a01db0f19b118a0c4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz - ## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /google-node-typescript-http-endpoint/package.json</p>
<p>Path to vulnerable library: /google-node-typescript-http-endpoint/node_modules/decode-uri-component/package.json,/aws-node-signed-uploads/node_modules/decode-uri-component/package.json,/azure-node-telegram-bot/node_modules/decode-uri-component/package.json,/aws-node-typescript-nest/node_modules/decode-uri-component/package.json,/aws-python-auth0-custom-authorizers-api/node_modules/decode-uri-component/package.json,/aws-rust-simple-http-endpoint/node_modules/decode-uri-component/package.json,/azure-node-typescript-servicebus-trigger-endpoint/node_modules/decode-uri-component/package.json,/aws-node-typescript-apollo-lambda/node_modules/decode-uri-component/package.json,/aws-node-fullstack/frontend/node_modules/decode-uri-component/package.json,/aws-node-dynamic-image-resizer/node_modules/decode-uri-component/package.json,/aws-node-rest-api-typescript-simple/node_modules/decode-uri-component/package.json,/azure-node-line-bot/node_modules/decode-uri-component/package.json,/aws-node-function-compiled-with-babel/node_modules/decode-uri-component/package.json,/aws-node-github-check/node_modules/decode-uri-component/package.json,/aws-node-typescript-sqs-standard/node_modules/decode-uri-component/package.json,/aws-node-vue-nuxt-ssr/node_modules/decode-uri-component/package.json,/aws-node-http-api-typescript/node_modules/decode-uri-component/package.json,/aws-node-typescript-kinesis/node_modules/decode-uri-component/package.json,/aws-golang-dynamo-stream-to-elasticsearch/node_modules/decode-uri-component/package.json,/aws-golang-auth-examples/node_modules/decode-uri-component/package.json,/aws-node-oauth-dropbox-api/node_modules/decode-uri-component/package.json,/aws-node-rest-api-typescript/node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- webpack-4.28.3.tgz
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- source-map-resolve-0.5.2.tgz
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/artkamote/examples/commit/dcbe4aefe4b3685f4b15493a01db0f19b118a0c4">dcbe4aefe4b3685f4b15493a01db0f19b118a0c4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in decode uri component tgz cve high severity vulnerability vulnerable library decode uri component tgz a better decodeuricomponent library home page a href path to dependency file google node typescript http endpoint package json path to vulnerable library google node typescript http endpoint node modules decode uri component package json aws node signed uploads node modules decode uri component package json azure node telegram bot node modules decode uri component package json aws node typescript nest node modules decode uri component package json aws python custom authorizers api node modules decode uri component package json aws rust simple http endpoint node modules decode uri component package json azure node typescript servicebus trigger endpoint node modules decode uri component package json aws node typescript apollo lambda node modules decode uri component package json aws node fullstack frontend node modules decode uri component package json aws node dynamic image resizer node modules decode uri component package json aws node rest api typescript simple node modules decode uri component package json azure node line bot node modules decode uri component package json aws node function compiled with babel node modules decode uri component package json aws node github check node modules decode uri component package json aws node typescript sqs standard node modules decode uri component package json aws node vue nuxt ssr node modules decode uri component package json aws node http api typescript node modules decode uri component package json aws node typescript kinesis node modules decode uri component package json aws golang dynamo stream to elasticsearch node modules decode uri component package json aws golang auth examples node modules decode uri component package json aws node oauth dropbox api node modules decode uri component package json aws node rest api typescript node modules decode uri component package json dependency hierarchy react scripts tgz root library webpack tgz micromatch tgz snapdragon tgz source map resolve tgz x decode uri component tgz vulnerable library found in head commit a href found in base branch master vulnerability details decode uri component is vulnerable to improper input validation resulting in dos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
44,720 | 18,168,491,374 | IssuesEvent | 2021-09-27 17:04:57 | boto/boto3 | https://api.github.com/repos/boto/boto3 | closed | Sudden Issue Overriding Batch Job Definition Resource Requirements | bug response-requested service-api | **Describe the bug**
Within the last week, the pipeline we use to submit jobs to AWS Batch via `boto3` client has broken, and attempting to follow the documentation is not producing a fix.
Specifically, we use the ` boto3` Batch client's `submit_job` method to submit jobs, wherein each job has it's own VCPU and memory requirements that are used to override the job definitions default values. These overrides are passed to the `containerOverrides` arg., traditionally using the `vcpus` and `memory` entries (e.g. `{'vcpus': 123,'memory': 123}`); however, now this produces the warning in the Batch console (and overrides the resource requirements):
> Deprecated key notice
This value was submitted using the vcpus key inside of containerProperties or containerOverrides, which has been deprecated. Instead, this value should be set inside the resourceRequirements key. More information about the deprecated key can be found in the AWS Batch API documentation. Learn more
Following the instructions given in the warning and at the "Learn more", I pass the `resourceRequirements` key to `containerOverrides`, instead of the two keypairs used above, as such:
```
{
'resourceRequirements': [
{'value': '15','type': 'VCPU'},
{'value': '200000','type': 'MEMORY'}
]
}
```
This, however, produces the following warning in Batch console:
> Configuration conflict
This value was submitted using containerOverrides.memory which has been deprecated and was not used as an override. Instead, the MEMORY value found in the job definition’s resourceRequirements key was used instead. More information about the deprecated key can be found in the AWS Batch API documentation. Learn more
Notably, *the default values from the job definition are used*.
We are using `boto3==1.18.44`.
**Steps to reproduce**
Using:
```{py}
import boto3
client = boto3.client(
'batch',
region_name=-REGION
)
```
Version 1:
- Results in warning but does override default values.
```{py}
container_overrides = {
'vcpus': 15,
'memory': 200000,
'command': 'bash sleep 30'.split(' '),
}
response = client.submit_job(
jobName=job_name,
jobQueue=job_queue,
jobDefinition=job_definition,
containerOverrides=container_overrides,
)
```
Version 2:
- Results in warning and does not override default values
```{py}
container_overrides = {
'resourceRequirements': [
{
'value': '15',
'type': 'VCPU'
},
{
'value': '200000',
'type': 'MEMORY'
},
]
'command': 'bash sleep 30'.split(' '),
}
response = client.submit_job(
jobName=job_name,
jobQueue=job_queue,
jobDefinition=job_definition,
containerOverrides=container_overrides,
)
```
**Expected behavior**
I would like to know the correct way to submit jobs that override the job definition resource requirements in a manner that does not produce an errors.
**Debug logs**
See screenshots for warnings.


| 1.0 | Sudden Issue Overriding Batch Job Definition Resource Requirements - **Describe the bug**
Within the last week, the pipeline we use to submit jobs to AWS Batch via `boto3` client has broken, and attempting to follow the documentation is not producing a fix.
Specifically, we use the ` boto3` Batch client's `submit_job` method to submit jobs, wherein each job has it's own VCPU and memory requirements that are used to override the job definitions default values. These overrides are passed to the `containerOverrides` arg., traditionally using the `vcpus` and `memory` entries (e.g. `{'vcpus': 123,'memory': 123}`); however, now this produces the warning in the Batch console (and overrides the resource requirements):
> Deprecated key notice
This value was submitted using the vcpus key inside of containerProperties or containerOverrides, which has been deprecated. Instead, this value should be set inside the resourceRequirements key. More information about the deprecated key can be found in the AWS Batch API documentation. Learn more
Following the instructions given in the warning and at the "Learn more", I pass the `resourceRequirements` key to `containerOverrides`, instead of the two keypairs used above, as such:
```
{
'resourceRequirements': [
{'value': '15','type': 'VCPU'},
{'value': '200000','type': 'MEMORY'}
]
}
```
This, however, produces the following warning in Batch console:
> Configuration conflict
This value was submitted using containerOverrides.memory which has been deprecated and was not used as an override. Instead, the MEMORY value found in the job definition’s resourceRequirements key was used instead. More information about the deprecated key can be found in the AWS Batch API documentation. Learn more
Notably, *the default values from the job definition are used*.
We are using `boto3==1.18.44`.
**Steps to reproduce**
Using:
```{py}
import boto3
client = boto3.client(
'batch',
region_name=-REGION
)
```
Version 1:
- Results in warning but does override default values.
```{py}
container_overrides = {
'vcpus': 15,
'memory': 200000,
'command': 'bash sleep 30'.split(' '),
}
response = client.submit_job(
jobName=job_name,
jobQueue=job_queue,
jobDefinition=job_definition,
containerOverrides=container_overrides,
)
```
Version 2:
- Results in warning and does not override default values
```{py}
container_overrides = {
'resourceRequirements': [
{
'value': '15',
'type': 'VCPU'
},
{
'value': '200000',
'type': 'MEMORY'
},
]
'command': 'bash sleep 30'.split(' '),
}
response = client.submit_job(
jobName=job_name,
jobQueue=job_queue,
jobDefinition=job_definition,
containerOverrides=container_overrides,
)
```
**Expected behavior**
I would like to know the correct way to submit jobs that override the job definition resource requirements in a manner that does not produce an errors.
**Debug logs**
See screenshots for warnings.


| non_defect | sudden issue overriding batch job definition resource requirements describe the bug within the last week the pipeline we use to submit jobs to aws batch via client has broken and attempting to follow the documentation is not producing a fix specifically we use the batch client s submit job method to submit jobs wherein each job has it s own vcpu and memory requirements that are used to override the job definitions default values these overrides are passed to the containeroverrides arg traditionally using the vcpus and memory entries e g vcpus memory however now this produces the warning in the batch console and overrides the resource requirements deprecated key notice this value was submitted using the vcpus key inside of containerproperties or containeroverrides which has been deprecated instead this value should be set inside the resourcerequirements key more information about the deprecated key can be found in the aws batch api documentation learn more following the instructions given in the warning and at the learn more i pass the resourcerequirements key to containeroverrides instead of the two keypairs used above as such resourcerequirements value type vcpu value type memory this however produces the following warning in batch console configuration conflict this value was submitted using containeroverrides memory which has been deprecated and was not used as an override instead the memory value found in the job definition’s resourcerequirements key was used instead more information about the deprecated key can be found in the aws batch api documentation learn more notably the default values from the job definition are used we are using steps to reproduce using py import client client batch region name region version results in warning but does override default values py container overrides vcpus memory command bash sleep split response client submit job jobname job name jobqueue job queue jobdefinition job definition containeroverrides container overrides version results in warning and does not override default values py container overrides resourcerequirements value type vcpu value type memory command bash sleep split response client submit job jobname job name jobqueue job queue jobdefinition job definition containeroverrides container overrides expected behavior i would like to know the correct way to submit jobs that override the job definition resource requirements in a manner that does not produce an errors debug logs see screenshots for warnings | 0 |
36,954 | 9,933,087,297 | IssuesEvent | 2019-07-02 11:27:45 | neovim/neovim | https://api.github.com/repos/neovim/neovim | closed | xfpclassify: warning: conversion from ‘double’ to ‘float’ may change value (gcc 9.1) | build | I am seeing the following warning, _likely_ caused by upgrading to gcc 9.1:
```
[226/303] Building C object src/nvim/CMakeFiles/nvim.dir/math.c.o
In file included from ../src/nvim/math.c:4:
../src/nvim/math.c: In function ‘xfpclassify’:
../src/nvim/math.c:24:21: warning: conversion from ‘double’ to ‘float’ may change value [-Wfloat-conversion]
24 | return fpclassify(d);
| ^
../src/nvim/math.c:24:21: warning: conversion from ‘double’ to ‘float’ may change value [-Wfloat-conversion]
24 | return fpclassify(d);
| ^
```
I've seen https://github.com/neovim/neovim/pull/9532, but have not investigated further yet. | 1.0 | xfpclassify: warning: conversion from ‘double’ to ‘float’ may change value (gcc 9.1) - I am seeing the following warning, _likely_ caused by upgrading to gcc 9.1:
```
[226/303] Building C object src/nvim/CMakeFiles/nvim.dir/math.c.o
In file included from ../src/nvim/math.c:4:
../src/nvim/math.c: In function ‘xfpclassify’:
../src/nvim/math.c:24:21: warning: conversion from ‘double’ to ‘float’ may change value [-Wfloat-conversion]
24 | return fpclassify(d);
| ^
../src/nvim/math.c:24:21: warning: conversion from ‘double’ to ‘float’ may change value [-Wfloat-conversion]
24 | return fpclassify(d);
| ^
```
I've seen https://github.com/neovim/neovim/pull/9532, but have not investigated further yet. | non_defect | xfpclassify warning conversion from ‘double’ to ‘float’ may change value gcc i am seeing the following warning likely caused by upgrading to gcc building c object src nvim cmakefiles nvim dir math c o in file included from src nvim math c src nvim math c in function ‘xfpclassify’ src nvim math c warning conversion from ‘double’ to ‘float’ may change value return fpclassify d src nvim math c warning conversion from ‘double’ to ‘float’ may change value return fpclassify d i ve seen but have not investigated further yet | 0 |
110,492 | 16,980,335,190 | IssuesEvent | 2021-06-30 08:02:06 | samuelteixeiras/learnRuby | https://api.github.com/repos/samuelteixeiras/learnRuby | opened | CVE-2012-3463 (Medium) detected in actionpack-3.0.9.gem | security vulnerability | ## CVE-2012-3463 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-3.0.9.gem</b></p></summary>
<p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionpack-3.0.9.gem">https://rubygems.org/gems/actionpack-3.0.9.gem</a></p>
<p>
Dependency Hierarchy:
- rails-3.0.9.gem (Root Library)
- railties-3.0.9.gem
- :x: **actionpack-3.0.9.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samuelteixeiras/learnRuby/commit/0c785fd9400921392b8ee5e3e166f30364359ecc">0c785fd9400921392b8ee5e3e166f30364359ecc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in actionpack/lib/action_view/helpers/form_tag_helper.rb in Ruby on Rails 3.x before 3.0.17, 3.1.x before 3.1.8, and 3.2.x before 3.2.8 allows remote attackers to inject arbitrary web script or HTML via the prompt field to the select_tag helper.
<p>Publish Date: 2012-08-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-3463>CVE-2012-3463</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-3463">https://nvd.nist.gov/vuln/detail/CVE-2012-3463</a></p>
<p>Release Date: 2012-08-10</p>
<p>Fix Resolution: 3.0.17,3.1.8,3.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2012-3463 (Medium) detected in actionpack-3.0.9.gem - ## CVE-2012-3463 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionpack-3.0.9.gem</b></p></summary>
<p>Web apps on Rails. Simple, battle-tested conventions for building and testing MVC web applications. Works with any Rack-compatible server.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionpack-3.0.9.gem">https://rubygems.org/gems/actionpack-3.0.9.gem</a></p>
<p>
Dependency Hierarchy:
- rails-3.0.9.gem (Root Library)
- railties-3.0.9.gem
- :x: **actionpack-3.0.9.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samuelteixeiras/learnRuby/commit/0c785fd9400921392b8ee5e3e166f30364359ecc">0c785fd9400921392b8ee5e3e166f30364359ecc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-site scripting (XSS) vulnerability in actionpack/lib/action_view/helpers/form_tag_helper.rb in Ruby on Rails 3.x before 3.0.17, 3.1.x before 3.1.8, and 3.2.x before 3.2.8 allows remote attackers to inject arbitrary web script or HTML via the prompt field to the select_tag helper.
<p>Publish Date: 2012-08-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-3463>CVE-2012-3463</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-3463">https://nvd.nist.gov/vuln/detail/CVE-2012-3463</a></p>
<p>Release Date: 2012-08-10</p>
<p>Fix Resolution: 3.0.17,3.1.8,3.2.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in actionpack gem cve medium severity vulnerability vulnerable library actionpack gem web apps on rails simple battle tested conventions for building and testing mvc web applications works with any rack compatible server library home page a href dependency hierarchy rails gem root library railties gem x actionpack gem vulnerable library found in head commit a href found in base branch master vulnerability details cross site scripting xss vulnerability in actionpack lib action view helpers form tag helper rb in ruby on rails x before x before and x before allows remote attackers to inject arbitrary web script or html via the prompt field to the select tag helper publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
82,056 | 31,884,481,503 | IssuesEvent | 2023-09-16 19:34:06 | nats-io/nats.py | https://api.github.com/repos/nats-io/nats.py | opened | Can't install v2.4.0 with Poetry | defect | ### What version were you using?
Client: v2.4.0
### What environment was the server running in?
Arch Linux x64
### Is this defect reproducible?
Yes: `$ poetry add nats-py`
### Given the capability you are leveraging, describe your expectation?
I should be able to install client v2.4.0, like I do with v2.3.1:
```
$ poetry install
Installing dependencies from lock file
Package operations: 1 install, 0 updates, 0 removals
• Installing nats-py (2.3.1)
```
### Given the expectation, what is the defect you are observing?
I can't install client v2.4.0:
```
$ poetry install
Installing dependencies from lock file
Package operations: 1 install, 0 updates, 0 removals
• Installing nats-py (2.4.0): Failed
TypeError
Can't instantiate abstract class IsolatedEnv with abstract methods make_extra_environ, python_executable
at /usr/lib/python3.11/site-packages/poetry/installation/chef.py:109 in _prepare
105│ ) -> Path:
106│ from subprocess import CalledProcessError
107│
108│ with ephemeral_environment(self._env.python) as venv:
→ 109│ env = IsolatedEnv(venv, self._pool)
110│ builder = ProjectBuilder(
111│ directory,
112│ python_executable=env.executable,
113│ scripts_dir=env.scripts_dir,
``` | 1.0 | Can't install v2.4.0 with Poetry - ### What version were you using?
Client: v2.4.0
### What environment was the server running in?
Arch Linux x64
### Is this defect reproducible?
Yes: `$ poetry add nats-py`
### Given the capability you are leveraging, describe your expectation?
I should be able to install client v2.4.0, like I do with v2.3.1:
```
$ poetry install
Installing dependencies from lock file
Package operations: 1 install, 0 updates, 0 removals
• Installing nats-py (2.3.1)
```
### Given the expectation, what is the defect you are observing?
I can't install client v2.4.0:
```
$ poetry install
Installing dependencies from lock file
Package operations: 1 install, 0 updates, 0 removals
• Installing nats-py (2.4.0): Failed
TypeError
Can't instantiate abstract class IsolatedEnv with abstract methods make_extra_environ, python_executable
at /usr/lib/python3.11/site-packages/poetry/installation/chef.py:109 in _prepare
105│ ) -> Path:
106│ from subprocess import CalledProcessError
107│
108│ with ephemeral_environment(self._env.python) as venv:
→ 109│ env = IsolatedEnv(venv, self._pool)
110│ builder = ProjectBuilder(
111│ directory,
112│ python_executable=env.executable,
113│ scripts_dir=env.scripts_dir,
``` | defect | can t install with poetry what version were you using client what environment was the server running in arch linux is this defect reproducible yes poetry add nats py given the capability you are leveraging describe your expectation i should be able to install client like i do with poetry install installing dependencies from lock file package operations install updates removals • installing nats py given the expectation what is the defect you are observing i can t install client poetry install installing dependencies from lock file package operations install updates removals • installing nats py failed typeerror can t instantiate abstract class isolatedenv with abstract methods make extra environ python executable at usr lib site packages poetry installation chef py in prepare │ path │ from subprocess import calledprocesserror │ │ with ephemeral environment self env python as venv → │ env isolatedenv venv self pool │ builder projectbuilder │ directory │ python executable env executable │ scripts dir env scripts dir | 1 |
7,175 | 2,610,355,593 | IssuesEvent | 2015-02-26 19:55:10 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | does not add a new blog for wordpress | auto-migrated Priority-Medium Type-Defect | ```
What's the problem?
scribefire does not add a new blog for wordpress
What browser are you using?
firefox 7.0.1
What version of ScribeFire are you running?
scribefire 1.9 tab
```
-----
Original issue reported on code.google.com by `theangry...@gmail.com` on 8 Oct 2011 at 3:21 | 1.0 | does not add a new blog for wordpress - ```
What's the problem?
scribefire does not add a new blog for wordpress
What browser are you using?
firefox 7.0.1
What version of ScribeFire are you running?
scribefire 1.9 tab
```
-----
Original issue reported on code.google.com by `theangry...@gmail.com` on 8 Oct 2011 at 3:21 | defect | does not add a new blog for wordpress what s the problem scribefire does not add a new blog for wordpress what browser are you using firefox what version of scribefire are you running scribefire tab original issue reported on code google com by theangry gmail com on oct at | 1 |
44,572 | 9,605,195,691 | IssuesEvent | 2019-05-10 22:44:44 | EdenServer/community | https://api.github.com/repos/EdenServer/community | closed | Mining Colored Rocks in Gusgen Mines issue | in-code-review | Colored Rocks are supposed to match the element of the day when mining them, e.g. firesday = red rock.
Currently this mostly works fine except Watersday and Earthsday's yields are switched. | 1.0 | Mining Colored Rocks in Gusgen Mines issue - Colored Rocks are supposed to match the element of the day when mining them, e.g. firesday = red rock.
Currently this mostly works fine except Watersday and Earthsday's yields are switched. | non_defect | mining colored rocks in gusgen mines issue colored rocks are supposed to match the element of the day when mining them e g firesday red rock currently this mostly works fine except watersday and earthsday s yields are switched | 0 |
22,498 | 3,787,585,503 | IssuesEvent | 2016-03-21 11:19:40 | ayooo-team/landing-page | https://api.github.com/repos/ayooo-team/landing-page | opened | Styling for Landing Page | design in-progress priority-1 T4hr | - [ ] simple navbar
- [ ] hero banner with temp unsplash image
- [ ] anchor tags
- [ ] flex-box (container-div) for 'how it works' and 'why ayooo' components
- [ ] divs for each image and text pair (component)
- [ ] large footer | 1.0 | Styling for Landing Page - - [ ] simple navbar
- [ ] hero banner with temp unsplash image
- [ ] anchor tags
- [ ] flex-box (container-div) for 'how it works' and 'why ayooo' components
- [ ] divs for each image and text pair (component)
- [ ] large footer | non_defect | styling for landing page simple navbar hero banner with temp unsplash image anchor tags flex box container div for how it works and why ayooo components divs for each image and text pair component large footer | 0 |
17,904 | 3,013,579,249 | IssuesEvent | 2015-07-29 09:50:16 | yawlfoundation/yawl | https://api.github.com/repos/yawlfoundation/yawl | closed | Custom icons not displayed correctly in the YAWl Editor 2.3 | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Create a 24x24 pixel png and move to icon directory for editor
2. Start Editor and create new specification
3. Icon is displayed correctly in the tollbar on the left but a flaming
"x-icon" is displayed in the specification.
What is the expected output? What do you see instead?
I would expect the same icon, taht is displayed in the toolbar on the left.
What version of the product are you using? On what operating system?
YAWL Editor 2.3 on Windows 7.
```
Original issue reported on code.google.com by `MarkusRo...@googlemail.com` on 30 Aug 2012 at 7:57
Attachments:
* [Editor Error.png](https://storage.googleapis.com/google-code-attachments/yawl/issue-469/comment-0/Editor Error.png)
| 1.0 | Custom icons not displayed correctly in the YAWl Editor 2.3 - ```
What steps will reproduce the problem?
1. Create a 24x24 pixel png and move to icon directory for editor
2. Start Editor and create new specification
3. Icon is displayed correctly in the tollbar on the left but a flaming
"x-icon" is displayed in the specification.
What is the expected output? What do you see instead?
I would expect the same icon, taht is displayed in the toolbar on the left.
What version of the product are you using? On what operating system?
YAWL Editor 2.3 on Windows 7.
```
Original issue reported on code.google.com by `MarkusRo...@googlemail.com` on 30 Aug 2012 at 7:57
Attachments:
* [Editor Error.png](https://storage.googleapis.com/google-code-attachments/yawl/issue-469/comment-0/Editor Error.png)
| defect | custom icons not displayed correctly in the yawl editor what steps will reproduce the problem create a pixel png and move to icon directory for editor start editor and create new specification icon is displayed correctly in the tollbar on the left but a flaming x icon is displayed in the specification what is the expected output what do you see instead i would expect the same icon taht is displayed in the toolbar on the left what version of the product are you using on what operating system yawl editor on windows original issue reported on code google com by markusro googlemail com on aug at attachments error png | 1 |
60,249 | 17,023,380,130 | IssuesEvent | 2021-07-03 01:43:35 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Problems with Help | Component: admin Priority: major Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 9.55am, Wednesday, 1st April 2009]**
Hi,
I use Version 1519(German).
When go to Hilfe/Hilfe (Help/Help) I come to a JOSM-Onlinehilfe Page with 2 Links. The link to the english Help-Site is OK, the other one produces an "Unerwarteter Fehler" (unexpected Error)
If I click on the "Im Browser ffnen" (Open in browser) Button I ended up on
<http://www.de.com//wiki.openstreetmap.org/wiki/DE:JOSM/Anleitung>
but this does not happens every time
When I press F1 Instead of going over the Menue, I get an empty page
| 1.0 | Problems with Help - **[Submitted to the original trac issue database at 9.55am, Wednesday, 1st April 2009]**
Hi,
I use Version 1519(German).
When go to Hilfe/Hilfe (Help/Help) I come to a JOSM-Onlinehilfe Page with 2 Links. The link to the english Help-Site is OK, the other one produces an "Unerwarteter Fehler" (unexpected Error)
If I click on the "Im Browser ffnen" (Open in browser) Button I ended up on
<http://www.de.com//wiki.openstreetmap.org/wiki/DE:JOSM/Anleitung>
but this does not happens every time
When I press F1 Instead of going over the Menue, I get an empty page
| defect | problems with help hi i use version german when go to hilfe hilfe help help i come to a josm onlinehilfe page with links the link to the english help site is ok the other one produces an unerwarteter fehler unexpected error if i click on the im browser ffnen open in browser button i ended up on but this does not happens every time when i press instead of going over the menue i get an empty page | 1 |
48,356 | 13,068,461,200 | IssuesEvent | 2020-07-31 03:39:07 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | [RFC] transitioning to python3 (Trac #2225) | Migrated from Trac cmake defect | Python 2.7 will not be maintained past 2020 (see https://pythonclock.org/ for details) other libraries we depend on will stop support earlier (such as numpy). We need a migration plan. Fortunately almost all of our python is compatible with python3. The use of python3 in cvmfs will make this pretty painless for the cluster computing environments, but there will be issues where this is not available such as user's laptops.
The way python is currently detected in cmake is that it checks the version of the interpreter by calling `python` and then links against the version of libboost_python which matches the version.
all of our executable python scripts the rely on the first line of each script starting with `#!/usr/bin/env python` to instantiate the python interpreter. It is very important that this interpreter be the same version of python as the version of libboost_python that icetray was compiled against. If cmake is run in an environment where `python` refers to python3 then icetray will automatically do everything right and work in python3. However, if the environment is such that `python` refers to python2 then trying to switch to python3 will cause problems.
Most systems by now have python3 but on most of them python2 is the default. There are two ways of ensuring that icetray uses python3 on a system where python2 is the default:
- First enter a python3 virtual environment before running cmake and reenter this same environment before running any icetray script.
- rewrite our cmake scripts and all of our scripts to explicitly call python3, this means changing every single hashbang including the users own scripts or always expliclty calling python3 from the command line.
I recommend the former, but the latter is possible. I also recommend that the next version of offline include warnings if it detects python2.
Proposed action items:
1. Write docs explaining how to migrate to python3 using virtualenv
2. Add waning to cmake's python detection if still using python2
3. add check to `env-shell.sh` to make sure the python interpreter in the path matches what was compiled against
4. add check in `load_pybindings.py` to make sure the the python version matches the what was compiled (the is mostly redundant as you are in a pretty messed environment if the check in `env-shell.sh` passed)
The python warnings can the be turned into a fatal errors in future releases.
Other suggestions on how to handle this are welcome.
Migrated from https://code.icecube.wisc.edu/ticket/2225
```json
{
"status": "closed",
"changetime": "2019-03-11T16:45:31",
"description": "Python 2.7 will not be maintained past 2020 (see https://pythonclock.org/ for details) other libraries we depend on will stop support earlier (such as numpy). We need a migration plan. Fortunately almost all of our python is compatible with python3. The use of python3 in cvmfs will make this pretty painless for the cluster computing environments, but there will be issues where this is not available such as user's laptops.\n\nThe way python is currently detected in cmake is that it checks the version of the interpreter by calling `python` and then links against the version of libboost_python which matches the version. \nall of our executable python scripts the rely on the first line of each script starting with `#!/usr/bin/env python` to instantiate the python interpreter. It is very important that this interpreter be the same version of python as the version of libboost_python that icetray was compiled against. If cmake is run in an environment where `python` refers to python3 then icetray will automatically do everything right and work in python3. However, if the environment is such that `python` refers to python2 then trying to switch to python3 will cause problems. \n\nMost systems by now have python3 but on most of them python2 is the default. There are two ways of ensuring that icetray uses python3 on a system where python2 is the default:\n- First enter a python3 virtual environment before running cmake and reenter this same environment before running any icetray script.\n- rewrite our cmake scripts and all of our scripts to explicitly call python3, this means changing every single hashbang including the users own scripts or always expliclty calling python3 from the command line.\n\nI recommend the former, but the latter is possible. I also recommend that the next version of offline include warnings if it detects python2. \n\n\nProposed action items:\n\n1. Write docs explaining how to migrate to python3 using virtualenv\n2. Add waning to cmake's python detection if still using python2\n3. add check to `env-shell.sh` to make sure the python interpreter in the path matches what was compiled against\n4. add check in `load_pybindings.py` to make sure the the python version matches the what was compiled (the is mostly redundant as you are in a pretty messed environment if the check in `env-shell.sh` passed)\n\nThe python warnings can the be turned into a fatal errors in future releases.\n\nOther suggestions on how to handle this are welcome. \n\n \n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1552322731408083",
"component": "cmake",
"summary": "[RFC] transitioning to python3",
"priority": "normal",
"keywords": "",
"time": "2018-12-13T17:08:51",
"milestone": "Vernal Equinox 2019",
"owner": "kjmeagher",
"type": "defect"
}
```
| 1.0 | [RFC] transitioning to python3 (Trac #2225) - Python 2.7 will not be maintained past 2020 (see https://pythonclock.org/ for details) other libraries we depend on will stop support earlier (such as numpy). We need a migration plan. Fortunately almost all of our python is compatible with python3. The use of python3 in cvmfs will make this pretty painless for the cluster computing environments, but there will be issues where this is not available such as user's laptops.
The way python is currently detected in cmake is that it checks the version of the interpreter by calling `python` and then links against the version of libboost_python which matches the version.
all of our executable python scripts the rely on the first line of each script starting with `#!/usr/bin/env python` to instantiate the python interpreter. It is very important that this interpreter be the same version of python as the version of libboost_python that icetray was compiled against. If cmake is run in an environment where `python` refers to python3 then icetray will automatically do everything right and work in python3. However, if the environment is such that `python` refers to python2 then trying to switch to python3 will cause problems.
Most systems by now have python3 but on most of them python2 is the default. There are two ways of ensuring that icetray uses python3 on a system where python2 is the default:
- First enter a python3 virtual environment before running cmake and reenter this same environment before running any icetray script.
- rewrite our cmake scripts and all of our scripts to explicitly call python3, this means changing every single hashbang including the users own scripts or always expliclty calling python3 from the command line.
I recommend the former, but the latter is possible. I also recommend that the next version of offline include warnings if it detects python2.
Proposed action items:
1. Write docs explaining how to migrate to python3 using virtualenv
2. Add waning to cmake's python detection if still using python2
3. add check to `env-shell.sh` to make sure the python interpreter in the path matches what was compiled against
4. add check in `load_pybindings.py` to make sure the the python version matches the what was compiled (the is mostly redundant as you are in a pretty messed environment if the check in `env-shell.sh` passed)
The python warnings can the be turned into a fatal errors in future releases.
Other suggestions on how to handle this are welcome.
Migrated from https://code.icecube.wisc.edu/ticket/2225
```json
{
"status": "closed",
"changetime": "2019-03-11T16:45:31",
"description": "Python 2.7 will not be maintained past 2020 (see https://pythonclock.org/ for details) other libraries we depend on will stop support earlier (such as numpy). We need a migration plan. Fortunately almost all of our python is compatible with python3. The use of python3 in cvmfs will make this pretty painless for the cluster computing environments, but there will be issues where this is not available such as user's laptops.\n\nThe way python is currently detected in cmake is that it checks the version of the interpreter by calling `python` and then links against the version of libboost_python which matches the version. \nall of our executable python scripts the rely on the first line of each script starting with `#!/usr/bin/env python` to instantiate the python interpreter. It is very important that this interpreter be the same version of python as the version of libboost_python that icetray was compiled against. If cmake is run in an environment where `python` refers to python3 then icetray will automatically do everything right and work in python3. However, if the environment is such that `python` refers to python2 then trying to switch to python3 will cause problems. \n\nMost systems by now have python3 but on most of them python2 is the default. There are two ways of ensuring that icetray uses python3 on a system where python2 is the default:\n- First enter a python3 virtual environment before running cmake and reenter this same environment before running any icetray script.\n- rewrite our cmake scripts and all of our scripts to explicitly call python3, this means changing every single hashbang including the users own scripts or always expliclty calling python3 from the command line.\n\nI recommend the former, but the latter is possible. I also recommend that the next version of offline include warnings if it detects python2. \n\n\nProposed action items:\n\n1. Write docs explaining how to migrate to python3 using virtualenv\n2. Add waning to cmake's python detection if still using python2\n3. add check to `env-shell.sh` to make sure the python interpreter in the path matches what was compiled against\n4. add check in `load_pybindings.py` to make sure the the python version matches the what was compiled (the is mostly redundant as you are in a pretty messed environment if the check in `env-shell.sh` passed)\n\nThe python warnings can the be turned into a fatal errors in future releases.\n\nOther suggestions on how to handle this are welcome. \n\n \n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1552322731408083",
"component": "cmake",
"summary": "[RFC] transitioning to python3",
"priority": "normal",
"keywords": "",
"time": "2018-12-13T17:08:51",
"milestone": "Vernal Equinox 2019",
"owner": "kjmeagher",
"type": "defect"
}
```
| defect | transitioning to trac python will not be maintained past see for details other libraries we depend on will stop support earlier such as numpy we need a migration plan fortunately almost all of our python is compatible with the use of in cvmfs will make this pretty painless for the cluster computing environments but there will be issues where this is not available such as user s laptops the way python is currently detected in cmake is that it checks the version of the interpreter by calling python and then links against the version of libboost python which matches the version all of our executable python scripts the rely on the first line of each script starting with usr bin env python to instantiate the python interpreter it is very important that this interpreter be the same version of python as the version of libboost python that icetray was compiled against if cmake is run in an environment where python refers to then icetray will automatically do everything right and work in however if the environment is such that python refers to then trying to switch to will cause problems most systems by now have but on most of them is the default there are two ways of ensuring that icetray uses on a system where is the default first enter a virtual environment before running cmake and reenter this same environment before running any icetray script rewrite our cmake scripts and all of our scripts to explicitly call this means changing every single hashbang including the users own scripts or always expliclty calling from the command line i recommend the former but the latter is possible i also recommend that the next version of offline include warnings if it detects proposed action items write docs explaining how to migrate to using virtualenv add waning to cmake s python detection if still using add check to env shell sh to make sure the python interpreter in the path matches what was compiled against add check in load pybindings py to make sure the the python version matches the what was compiled the is mostly redundant as you are in a pretty messed environment if the check in env shell sh passed the python warnings can the be turned into a fatal errors in future releases other suggestions on how to handle this are welcome migrated from json status closed changetime description python will not be maintained past see for details other libraries we depend on will stop support earlier such as numpy we need a migration plan fortunately almost all of our python is compatible with the use of in cvmfs will make this pretty painless for the cluster computing environments but there will be issues where this is not available such as user s laptops n nthe way python is currently detected in cmake is that it checks the version of the interpreter by calling python and then links against the version of libboost python which matches the version nall of our executable python scripts the rely on the first line of each script starting with usr bin env python to instantiate the python interpreter it is very important that this interpreter be the same version of python as the version of libboost python that icetray was compiled against if cmake is run in an environment where python refers to then icetray will automatically do everything right and work in however if the environment is such that python refers to then trying to switch to will cause problems n nmost systems by now have but on most of them is the default there are two ways of ensuring that icetray uses on a system where is the default n first enter a virtual environment before running cmake and reenter this same environment before running any icetray script n rewrite our cmake scripts and all of our scripts to explicitly call this means changing every single hashbang including the users own scripts or always expliclty calling from the command line n ni recommend the former but the latter is possible i also recommend that the next version of offline include warnings if it detects n n nproposed action items n write docs explaining how to migrate to using virtualenv add waning to cmake s python detection if still using add check to env shell sh to make sure the python interpreter in the path matches what was compiled against add check in load pybindings py to make sure the the python version matches the what was compiled the is mostly redundant as you are in a pretty messed environment if the check in env shell sh passed n nthe python warnings can the be turned into a fatal errors in future releases n nother suggestions on how to handle this are welcome n n n reporter kjmeagher cc resolution fixed ts component cmake summary transitioning to priority normal keywords time milestone vernal equinox owner kjmeagher type defect | 1 |
41,267 | 10,349,743,410 | IssuesEvent | 2019-09-04 23:43:48 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Parser missing type information for Enum fields | T: Defect | ### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve):
- Parse any schema having a table with an ENUM e.g., from the standard `employees` MySql sample db:
```sql
CREATE TABLE employees (
emp_no INT NOT NULL,
birth_date DATE NOT NULL,
first_name VARCHAR(14) NOT NULL,
last_name VARCHAR(16) NOT NULL,
gender ENUM ('M','F') NOT NULL,
hire_date DATE NOT NULL,
PRIMARY KEY (emp_no)
);
```
- Parse the schema:
```java
...
DSLContext ctx = DSL.using(SQLDialect.MYSQL);
Meta meta = ctx.meta(Source.of(employees));
```
- Use a debugger to examine the `Table` object corresponding with `employees` table. Notice there is no type information available for the GENDER field other than the `Object` data type. I expected to see type info for the enum constants:

### Versions:
- jOOQ: 3.12.0
- Java: 8
- JDBC Driver (include name if inofficial driver): H2
| 1.0 | Parser missing type information for Enum fields - ### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve):
- Parse any schema having a table with an ENUM e.g., from the standard `employees` MySql sample db:
```sql
CREATE TABLE employees (
emp_no INT NOT NULL,
birth_date DATE NOT NULL,
first_name VARCHAR(14) NOT NULL,
last_name VARCHAR(16) NOT NULL,
gender ENUM ('M','F') NOT NULL,
hire_date DATE NOT NULL,
PRIMARY KEY (emp_no)
);
```
- Parse the schema:
```java
...
DSLContext ctx = DSL.using(SQLDialect.MYSQL);
Meta meta = ctx.meta(Source.of(employees));
```
- Use a debugger to examine the `Table` object corresponding with `employees` table. Notice there is no type information available for the GENDER field other than the `Object` data type. I expected to see type info for the enum constants:

### Versions:
- jOOQ: 3.12.0
- Java: 8
- JDBC Driver (include name if inofficial driver): H2
| defect | parser missing type information for enum fields steps to reproduce the problem if possible create an mcve parse any schema having a table with an enum e g from the standard employees mysql sample db sql create table employees emp no int not null birth date date not null first name varchar not null last name varchar not null gender enum m f not null hire date date not null primary key emp no parse the schema java dslcontext ctx dsl using sqldialect mysql meta meta ctx meta source of employees use a debugger to examine the table object corresponding with employees table notice there is no type information available for the gender field other than the object data type i expected to see type info for the enum constants versions jooq java jdbc driver include name if inofficial driver | 1 |
46,706 | 13,055,962,192 | IssuesEvent | 2020-07-30 03:14:50 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | pybindings from frame_object_diff confuse sphinx (Trac #1741) | Incomplete Migration Migrated from Trac combo core defect | Migrated from https://code.icecube.wisc.edu/ticket/1741
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"description": "For many of the pybinding functions in frame_object_diff, sphinx complains that it cannot format the arguments. \n{{{\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3CalibrationDiff.__copy__: <Boost.Python.function object at 0x7fd17ddc5cd0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3CalibrationDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17ddc5d20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3CalibrationDiff.unpack: <Boost.Python.function object at 0x7fd17dcf26a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationDiff.__copy__: <Boost.Python.function object at 0x7fd17dcf6ed0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dcf6f20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationDiff.unpack: <Boost.Python.function object at 0x7fd17ddbdb70> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5a860> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5a8b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.plus_keys: <Boost.Python.function object at 0x7fd17dd5a7c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.plus_values: <Boost.Python.function object at 0x7fd17dd5a810> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.size: <Boost.Python.function object at 0x7fd17dd5a770> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.unpack: <Boost.Python.function object at 0x7fd17dd5a720> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5af00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5b250> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.append: <Boost.Python.function object at 0x7fd17dd5b5c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.extend: <Boost.Python.function object at 0x7fd17dd5b610> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DetectorStatusDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5bfe0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DetectorStatusDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5c030> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DetectorStatusDiff.unpack: <Boost.Python.function object at 0x7fd17dd5b1f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorIntDiff.__copy__: <Boost.Python.function object at 0x7fd17ddcd2e0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorIntDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17ddcd330> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorIntDiff.unpack: <Boost.Python.function object at 0x7fd17ddcd240> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorOMKeyDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5cbc0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorOMKeyDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5cc10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorOMKeyDiff.unpack: <Boost.Python.function object at 0x7fd17dd5cb20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorTankKeyDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5d7a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorTankKeyDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5d7f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorTankKeyDiff.unpack: <Boost.Python.function object at 0x7fd17dd5d700> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.__getitem__: <Boost.Python.function object at 0x7fd17dd9ec30> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.__setitem__: <Boost.Python.function object at 0x7fd17dd9ec80> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.reset: <Boost.Python.function object at 0x7fd17dd9ebe0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.set: <Boost.Python.function object at 0x7fd17dd9eb90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.__getitem__: <Boost.Python.function object at 0x7fd17dd9f5c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.__setitem__: <Boost.Python.function object at 0x7fd17dd9f610> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.reset: <Boost.Python.function object at 0x7fd17dd9f570> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.set: <Boost.Python.function object at 0x7fd17dd9f520> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.__getitem__: <Boost.Python.function object at 0x7fd17dd9ff50> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.__setitem__: <Boost.Python.function object at 0x7fd17dd9ffa0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.reset: <Boost.Python.function object at 0x7fd17dd9ff00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.set: <Boost.Python.function object at 0x7fd17dd9feb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.__getitem__: <Boost.Python.function object at 0x7fd17dda0260> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.__setitem__: <Boost.Python.function object at 0x7fd17ddf7870> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.reset: <Boost.Python.function object at 0x7fd17dda0210> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.set: <Boost.Python.function object at 0x7fd17dda01c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.__getitem__: <Boost.Python.function object at 0x7fd17ddf8230> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.__setitem__: <Boost.Python.function object at 0x7fd17ddf8280> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.reset: <Boost.Python.function object at 0x7fd17ddf81e0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.set: <Boost.Python.function object at 0x7fd17ddf8190> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.__getitem__: <Boost.Python.function object at 0x7fd17ddf8c40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.__setitem__: <Boost.Python.function object at 0x7fd17ddf8c90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.reset: <Boost.Python.function object at 0x7fd17ddf8bf0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.set: <Boost.Python.function object at 0x7fd17ddf8ba0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.__getitem__: <Boost.Python.function object at 0x7fd17dda6250> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.__setitem__: <Boost.Python.function object at 0x7fd17dda62a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.reset: <Boost.Python.function object at 0x7fd17dda6200> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.set: <Boost.Python.function object at 0x7fd17dda61b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.__getitem__: <Boost.Python.function object at 0x7fd17dda6c60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.__setitem__: <Boost.Python.function object at 0x7fd17dda6cb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.reset: <Boost.Python.function object at 0x7fd17dda6c10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.set: <Boost.Python.function object at 0x7fd17dda6bc0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.__getitem__: <Boost.Python.function object at 0x7fd17dda7670> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.__setitem__: <Boost.Python.function object at 0x7fd17dda76c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.reset: <Boost.Python.function object at 0x7fd17dda7620> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.set: <Boost.Python.function object at 0x7fd17dda75d0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.__getitem__: <Boost.Python.function object at 0x7fd17dda8080> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.__setitem__: <Boost.Python.function object at 0x7fd17dda80d0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.reset: <Boost.Python.function object at 0x7fd17dda8030> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.set: <Boost.Python.function object at 0x7fd17dda7fe0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.__getitem__: <Boost.Python.function object at 0x7fd17ddcddc0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.__setitem__: <Boost.Python.function object at 0x7fd17ddcde10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.reset: <Boost.Python.function object at 0x7fd17ddcdd70> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.set: <Boost.Python.function object at 0x7fd17ddcdd20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.__getitem__: <Boost.Python.function object at 0x7fd17dda8a90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.__setitem__: <Boost.Python.function object at 0x7fd17dda8ae0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.reset: <Boost.Python.function object at 0x7fd17dda8a40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.set: <Boost.Python.function object at 0x7fd17dda89f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.__getitem__: <Boost.Python.function object at 0x7fd17dda94a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.__setitem__: <Boost.Python.function object at 0x7fd17dda94f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.reset: <Boost.Python.function object at 0x7fd17dda9450> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.set: <Boost.Python.function object at 0x7fd17dda9400> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.__getitem__: <Boost.Python.function object at 0x7fd17dda9eb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.__setitem__: <Boost.Python.function object at 0x7fd17dda9f00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.reset: <Boost.Python.function object at 0x7fd17dda9e60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.set: <Boost.Python.function object at 0x7fd17dda9e10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.__getitem__: <Boost.Python.function object at 0x7fd17ddaa8c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.__setitem__: <Boost.Python.function object at 0x7fd17ddaa910> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.reset: <Boost.Python.function object at 0x7fd17ddaa870> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.set: <Boost.Python.function object at 0x7fd17ddaa820> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.__getitem__: <Boost.Python.function object at 0x7fd17dd348a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.__setitem__: <Boost.Python.function object at 0x7fd17dd348f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.reset: <Boost.Python.function object at 0x7fd17dd34850> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.set: <Boost.Python.function object at 0x7fd17dd34800> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.__getitem__: <Boost.Python.function object at 0x7fd17ddfa700> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.__setitem__: <Boost.Python.function object at 0x7fd17ddfa750> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.reset: <Boost.Python.function object at 0x7fd17ddfa6b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.set: <Boost.Python.function object at 0x7fd17ddfa660> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.__getitem__: <Boost.Python.function object at 0x7fd17ddfb0f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.__setitem__: <Boost.Python.function object at 0x7fd17ddfb140> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.reset: <Boost.Python.function object at 0x7fd17ddfb0a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.set: <Boost.Python.function object at 0x7fd17ddfb050> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.__getitem__: <Boost.Python.function object at 0x7fd17ddfbb00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.__setitem__: <Boost.Python.function object at 0x7fd17ddfbb50> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.reset: <Boost.Python.function object at 0x7fd17ddfbab0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.set: <Boost.Python.function object at 0x7fd17ddfba60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.__getitem__: <Boost.Python.function object at 0x7fd17dcae7f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.__setitem__: <Boost.Python.function object at 0x7fd17dcae840> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.reset: <Boost.Python.function object at 0x7fd17dcae7a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.set: <Boost.Python.function object at 0x7fd17dcae750> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.__getitem__: <Boost.Python.function object at 0x7fd17dd37bf0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.__setitem__: <Boost.Python.function object at 0x7fd17dd37c40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.reset: <Boost.Python.function object at 0x7fd17dd37ba0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.set: <Boost.Python.function object at 0x7fd17dd37b50> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.__getitem__: <Boost.Python.function object at 0x7fd17dd424b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.__setitem__: <Boost.Python.function object at 0x7fd17dd42500> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.reset: <Boost.Python.function object at 0x7fd17dd42460> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.set: <Boost.Python.function object at 0x7fd17dd42410> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.__getitem__: <Boost.Python.function object at 0x7fd17ddd7b40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.__setitem__: <Boost.Python.function object at 0x7fd17ddd7b90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.reset: <Boost.Python.function object at 0x7fd17ddd7af0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.set: <Boost.Python.function object at 0x7fd17ddd7aa0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.__getitem__: <Boost.Python.function object at 0x7fd17ddd8520> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.__setitem__: <Boost.Python.function object at 0x7fd17ddd8570> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.reset: <Boost.Python.function object at 0x7fd17ddd84d0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.set: <Boost.Python.function object at 0x7fd17ddd8480> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.__getitem__: <Boost.Python.function object at 0x7fd17ddd8f30> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.__setitem__: <Boost.Python.function object at 0x7fd17ddd8f80> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.reset: <Boost.Python.function object at 0x7fd17ddd8ee0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.set: <Boost.Python.function object at 0x7fd17ddd8e90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.__getitem__: <Boost.Python.function object at 0x7fd17dd43460> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.__setitem__: <Boost.Python.function object at 0x7fd17dd434b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.reset: <Boost.Python.function object at 0x7fd17dd43410> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.set: <Boost.Python.function object at 0x7fd17dd42da0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.__getitem__: <Boost.Python.function object at 0x7fd17dd48f60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.__setitem__: <Boost.Python.function object at 0x7fd17dd48fb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.reset: <Boost.Python.function object at 0x7fd17dd48f10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.set: <Boost.Python.function object at 0x7fd17dd48ec0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.__getitem__: <Boost.Python.function object at 0x7fd17dd43b40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.__setitem__: <Boost.Python.function object at 0x7fd17dd43b90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.reset: <Boost.Python.function object at 0x7fd17dd43af0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.set: <Boost.Python.function object at 0x7fd17dd43aa0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.__getitem__: <Boost.Python.function object at 0x7fd17dd26dd0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.__setitem__: <Boost.Python.function object at 0x7fd17dd26e20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.reset: <Boost.Python.function object at 0x7fd17dd26d80> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.set: <Boost.Python.function object at 0x7fd17dd26d30> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.__getitem__: <Boost.Python.function object at 0x7fd17dde7130> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.__setitem__: <Boost.Python.function object at 0x7fd17dde7180> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.reset: <Boost.Python.function object at 0x7fd17dde70e0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.set: <Boost.Python.function object at 0x7fd17dd273c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.__getitem__: <Boost.Python.function object at 0x7fd17dd9e2a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.__setitem__: <Boost.Python.function object at 0x7fd17dd9e2f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.reset: <Boost.Python.function object at 0x7fd17dd9e250> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.set: <Boost.Python.function object at 0x7fd17dd9e200> is not a Python function\n}}}",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067158057333",
"component": "combo core",
"summary": "pybindings from frame_object_diff confuse sphinx",
"priority": "minor",
"keywords": "documentation pybindings",
"time": "2016-06-10T11:15:02",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
| 1.0 | pybindings from frame_object_diff confuse sphinx (Trac #1741) - Migrated from https://code.icecube.wisc.edu/ticket/1741
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"description": "For many of the pybinding functions in frame_object_diff, sphinx complains that it cannot format the arguments. \n{{{\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3CalibrationDiff.__copy__: <Boost.Python.function object at 0x7fd17ddc5cd0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3CalibrationDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17ddc5d20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3CalibrationDiff.unpack: <Boost.Python.function object at 0x7fd17dcf26a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationDiff.__copy__: <Boost.Python.function object at 0x7fd17dcf6ed0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dcf6f20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationDiff.unpack: <Boost.Python.function object at 0x7fd17ddbdb70> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5a860> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5a8b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.plus_keys: <Boost.Python.function object at 0x7fd17dd5a7c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.plus_values: <Boost.Python.function object at 0x7fd17dd5a810> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.size: <Boost.Python.function object at 0x7fd17dd5a770> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationMapDiff.unpack: <Boost.Python.function object at 0x7fd17dd5a720> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5af00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5b250> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.append: <Boost.Python.function object at 0x7fd17dd5b5c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DOMCalibrationVectorDiff.extend: <Boost.Python.function object at 0x7fd17dd5b610> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DetectorStatusDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5bfe0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DetectorStatusDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5c030> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3DetectorStatusDiff.unpack: <Boost.Python.function object at 0x7fd17dd5b1f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorIntDiff.__copy__: <Boost.Python.function object at 0x7fd17ddcd2e0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorIntDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17ddcd330> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorIntDiff.unpack: <Boost.Python.function object at 0x7fd17ddcd240> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorOMKeyDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5cbc0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorOMKeyDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5cc10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorOMKeyDiff.unpack: <Boost.Python.function object at 0x7fd17dd5cb20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorTankKeyDiff.__copy__: <Boost.Python.function object at 0x7fd17dd5d7a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorTankKeyDiff.__deepcopy__: <Boost.Python.function object at 0x7fd17dd5d7f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.I3VectorTankKeyDiff.unpack: <Boost.Python.function object at 0x7fd17dd5d700> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.__getitem__: <Boost.Python.function object at 0x7fd17dd9ec30> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.__setitem__: <Boost.Python.function object at 0x7fd17dd9ec80> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.reset: <Boost.Python.function object at 0x7fd17dd9ebe0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset10.set: <Boost.Python.function object at 0x7fd17dd9eb90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.__getitem__: <Boost.Python.function object at 0x7fd17dd9f5c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.__setitem__: <Boost.Python.function object at 0x7fd17dd9f610> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.reset: <Boost.Python.function object at 0x7fd17dd9f570> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset11.set: <Boost.Python.function object at 0x7fd17dd9f520> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.__getitem__: <Boost.Python.function object at 0x7fd17dd9ff50> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.__setitem__: <Boost.Python.function object at 0x7fd17dd9ffa0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.reset: <Boost.Python.function object at 0x7fd17dd9ff00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset12.set: <Boost.Python.function object at 0x7fd17dd9feb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.__getitem__: <Boost.Python.function object at 0x7fd17dda0260> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.__setitem__: <Boost.Python.function object at 0x7fd17ddf7870> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.reset: <Boost.Python.function object at 0x7fd17dda0210> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset13.set: <Boost.Python.function object at 0x7fd17dda01c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.__getitem__: <Boost.Python.function object at 0x7fd17ddf8230> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.__setitem__: <Boost.Python.function object at 0x7fd17ddf8280> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.reset: <Boost.Python.function object at 0x7fd17ddf81e0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset14.set: <Boost.Python.function object at 0x7fd17ddf8190> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.__getitem__: <Boost.Python.function object at 0x7fd17ddf8c40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.__setitem__: <Boost.Python.function object at 0x7fd17ddf8c90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.reset: <Boost.Python.function object at 0x7fd17ddf8bf0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset15.set: <Boost.Python.function object at 0x7fd17ddf8ba0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.__getitem__: <Boost.Python.function object at 0x7fd17dda6250> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.__setitem__: <Boost.Python.function object at 0x7fd17dda62a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.reset: <Boost.Python.function object at 0x7fd17dda6200> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset16.set: <Boost.Python.function object at 0x7fd17dda61b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.__getitem__: <Boost.Python.function object at 0x7fd17dda6c60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.__setitem__: <Boost.Python.function object at 0x7fd17dda6cb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.reset: <Boost.Python.function object at 0x7fd17dda6c10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset17.set: <Boost.Python.function object at 0x7fd17dda6bc0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.__getitem__: <Boost.Python.function object at 0x7fd17dda7670> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.__setitem__: <Boost.Python.function object at 0x7fd17dda76c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.reset: <Boost.Python.function object at 0x7fd17dda7620> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset18.set: <Boost.Python.function object at 0x7fd17dda75d0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.__getitem__: <Boost.Python.function object at 0x7fd17dda8080> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.__setitem__: <Boost.Python.function object at 0x7fd17dda80d0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.reset: <Boost.Python.function object at 0x7fd17dda8030> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset19.set: <Boost.Python.function object at 0x7fd17dda7fe0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.__getitem__: <Boost.Python.function object at 0x7fd17ddcddc0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.__setitem__: <Boost.Python.function object at 0x7fd17ddcde10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.reset: <Boost.Python.function object at 0x7fd17ddcdd70> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset2.set: <Boost.Python.function object at 0x7fd17ddcdd20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.__getitem__: <Boost.Python.function object at 0x7fd17dda8a90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.__setitem__: <Boost.Python.function object at 0x7fd17dda8ae0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.reset: <Boost.Python.function object at 0x7fd17dda8a40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset20.set: <Boost.Python.function object at 0x7fd17dda89f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.__getitem__: <Boost.Python.function object at 0x7fd17dda94a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.__setitem__: <Boost.Python.function object at 0x7fd17dda94f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.reset: <Boost.Python.function object at 0x7fd17dda9450> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset21.set: <Boost.Python.function object at 0x7fd17dda9400> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.__getitem__: <Boost.Python.function object at 0x7fd17dda9eb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.__setitem__: <Boost.Python.function object at 0x7fd17dda9f00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.reset: <Boost.Python.function object at 0x7fd17dda9e60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset22.set: <Boost.Python.function object at 0x7fd17dda9e10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.__getitem__: <Boost.Python.function object at 0x7fd17ddaa8c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.__setitem__: <Boost.Python.function object at 0x7fd17ddaa910> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.reset: <Boost.Python.function object at 0x7fd17ddaa870> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset23.set: <Boost.Python.function object at 0x7fd17ddaa820> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.__getitem__: <Boost.Python.function object at 0x7fd17dd348a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.__setitem__: <Boost.Python.function object at 0x7fd17dd348f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.reset: <Boost.Python.function object at 0x7fd17dd34850> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset24.set: <Boost.Python.function object at 0x7fd17dd34800> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.__getitem__: <Boost.Python.function object at 0x7fd17ddfa700> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.__setitem__: <Boost.Python.function object at 0x7fd17ddfa750> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.reset: <Boost.Python.function object at 0x7fd17ddfa6b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset25.set: <Boost.Python.function object at 0x7fd17ddfa660> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.__getitem__: <Boost.Python.function object at 0x7fd17ddfb0f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.__setitem__: <Boost.Python.function object at 0x7fd17ddfb140> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.reset: <Boost.Python.function object at 0x7fd17ddfb0a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset26.set: <Boost.Python.function object at 0x7fd17ddfb050> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.__getitem__: <Boost.Python.function object at 0x7fd17ddfbb00> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.__setitem__: <Boost.Python.function object at 0x7fd17ddfbb50> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.reset: <Boost.Python.function object at 0x7fd17ddfbab0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset27.set: <Boost.Python.function object at 0x7fd17ddfba60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.__getitem__: <Boost.Python.function object at 0x7fd17dcae7f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.__setitem__: <Boost.Python.function object at 0x7fd17dcae840> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.reset: <Boost.Python.function object at 0x7fd17dcae7a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset28.set: <Boost.Python.function object at 0x7fd17dcae750> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.__getitem__: <Boost.Python.function object at 0x7fd17dd37bf0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.__setitem__: <Boost.Python.function object at 0x7fd17dd37c40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.reset: <Boost.Python.function object at 0x7fd17dd37ba0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset29.set: <Boost.Python.function object at 0x7fd17dd37b50> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.__getitem__: <Boost.Python.function object at 0x7fd17dd424b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.__setitem__: <Boost.Python.function object at 0x7fd17dd42500> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.reset: <Boost.Python.function object at 0x7fd17dd42460> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset3.set: <Boost.Python.function object at 0x7fd17dd42410> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.__getitem__: <Boost.Python.function object at 0x7fd17ddd7b40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.__setitem__: <Boost.Python.function object at 0x7fd17ddd7b90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.reset: <Boost.Python.function object at 0x7fd17ddd7af0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset30.set: <Boost.Python.function object at 0x7fd17ddd7aa0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.__getitem__: <Boost.Python.function object at 0x7fd17ddd8520> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.__setitem__: <Boost.Python.function object at 0x7fd17ddd8570> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.reset: <Boost.Python.function object at 0x7fd17ddd84d0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset31.set: <Boost.Python.function object at 0x7fd17ddd8480> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.__getitem__: <Boost.Python.function object at 0x7fd17ddd8f30> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.__setitem__: <Boost.Python.function object at 0x7fd17ddd8f80> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.reset: <Boost.Python.function object at 0x7fd17ddd8ee0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset32.set: <Boost.Python.function object at 0x7fd17ddd8e90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.__getitem__: <Boost.Python.function object at 0x7fd17dd43460> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.__setitem__: <Boost.Python.function object at 0x7fd17dd434b0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.reset: <Boost.Python.function object at 0x7fd17dd43410> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset4.set: <Boost.Python.function object at 0x7fd17dd42da0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.__getitem__: <Boost.Python.function object at 0x7fd17dd48f60> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.__setitem__: <Boost.Python.function object at 0x7fd17dd48fb0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.reset: <Boost.Python.function object at 0x7fd17dd48f10> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset5.set: <Boost.Python.function object at 0x7fd17dd48ec0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.__getitem__: <Boost.Python.function object at 0x7fd17dd43b40> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.__setitem__: <Boost.Python.function object at 0x7fd17dd43b90> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.reset: <Boost.Python.function object at 0x7fd17dd43af0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset6.set: <Boost.Python.function object at 0x7fd17dd43aa0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.__getitem__: <Boost.Python.function object at 0x7fd17dd26dd0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.__setitem__: <Boost.Python.function object at 0x7fd17dd26e20> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.reset: <Boost.Python.function object at 0x7fd17dd26d80> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset7.set: <Boost.Python.function object at 0x7fd17dd26d30> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.__getitem__: <Boost.Python.function object at 0x7fd17dde7130> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.__setitem__: <Boost.Python.function object at 0x7fd17dde7180> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.reset: <Boost.Python.function object at 0x7fd17dde70e0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset8.set: <Boost.Python.function object at 0x7fd17dd273c0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.__getitem__: <Boost.Python.function object at 0x7fd17dd9e2a0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.__setitem__: <Boost.Python.function object at 0x7fd17dd9e2f0> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.reset: <Boost.Python.function object at 0x7fd17dd9e250> is not a Python function\n/Users/kmeagher/icecube/combo/release/sphinx_build/source/python/icecube.frame_object_diff.rst:4: WARNING: error while formatting arguments for icecube.frame_object_diff.bitset9.set: <Boost.Python.function object at 0x7fd17dd9e200> is not a Python function\n}}}",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067158057333",
"component": "combo core",
"summary": "pybindings from frame_object_diff confuse sphinx",
"priority": "minor",
"keywords": "documentation pybindings",
"time": "2016-06-10T11:15:02",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
| defect | pybindings from frame object diff confuse sphinx trac migrated from json status closed changetime description for many of the pybinding functions in frame object diff sphinx complains that it cannot format the arguments n n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff unpack is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff unpack is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff plus keys is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff plus values is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff size is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff unpack is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff append is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff extend is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff unpack is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff unpack is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff unpack is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff copy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff deepcopy is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff unpack is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff getitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff setitem is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff reset is not a python function n users kmeagher icecube combo release sphinx build source python icecube frame object diff rst warning error while formatting arguments for icecube frame object diff set is not a python function n reporter kjmeagher cc resolution fixed ts component combo core summary pybindings from frame object diff confuse sphinx priority minor keywords documentation pybindings time milestone owner david schultz type defect | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.