Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
20,579
| 3,385,462,255
|
IssuesEvent
|
2015-11-27 11:39:31
|
jOOQ/jOOX
|
https://api.github.com/repos/jOOQ/jOOX
|
closed
|
Trouble writing CDATA. Getting HTML entity name instead of < and >
|
P: Medium T: defect
|
Never had trouble writing any content inside any tag, but now I'm facing this and I wouldn't like to call String.replace() before saving the XML file.
When Impl.text(Content content) is called, my value is "<![CDATA[ abc ]]>". Value is setted through match.setTextContent(text).
Before the method finishes, using debug, I saw the value of this.toString() and it is returning < and >. But if I call this.get(0), it returns my original value(having < and >).
I don't know if it is working as intended, maybe it is and I must look somewhere else.
Should I use a different approach or is there any configuration?
|
1.0
|
Trouble writing CDATA. Getting HTML entity name instead of < and > - Never had trouble writing any content inside any tag, but now I'm facing this and I wouldn't like to call String.replace() before saving the XML file.
When Impl.text(Content content) is called, my value is "<![CDATA[ abc ]]>". Value is setted through match.setTextContent(text).
Before the method finishes, using debug, I saw the value of this.toString() and it is returning < and >. But if I call this.get(0), it returns my original value(having < and >).
I don't know if it is working as intended, maybe it is and I must look somewhere else.
Should I use a different approach or is there any configuration?
|
defect
|
trouble writing cdata getting html entity name instead of never had trouble writing any content inside any tag but now i m facing this and i wouldn t like to call string replace before saving the xml file when impl text content content is called my value is value is setted through match settextcontent text before the method finishes using debug i saw the value of this tostring and it is returning lt and gt but if i call this get it returns my original value having i don t know if it is working as intended maybe it is and i must look somewhere else should i use a different approach or is there any configuration
| 1
|
327,335
| 24,128,753,784
|
IssuesEvent
|
2022-09-21 04:50:24
|
binotaliu/goodmoments.tw
|
https://api.github.com/repos/binotaliu/goodmoments.tw
|
closed
|
User Manual
|
documentation
|
User Manual related issues:
## Account
- [x] How to add an account
- [x] How to log in
- [x] How to disable an account
- [x] How to change/reset my password
- [ ] How to change/reset other person's password (if you're super administrator)
### FAQ
- [x] What is a good password
- [x] What to do if I have been told that my password has been in a data leak
- [x] What to do if I forgot my password
- [x] What to do if I can't log in
## Catalog
- [x] Good / Bad examples of product / category image
- [x] How to add a category
- [x] How to edit a category
- [x] How to remove a category
- [x] How to add a product
- [x] How to edit a product
- [x] How to remove a product
### FAQ
- [ ] What to do if I can't remove a category: category is not empty
- [x] What to do if I can't upload an image to category or product
- [ ] What to do if the syntax of product description is a mess
- [ ] What to do if I can't upload my image: file size too large
## Campaign
- [x] Good / Bad examples of banner image
- [x] How to add image to home page banner
- [x] How to schedule a banner item
- [x] How to remove an item from home page banner
- [x] Social media best practice
- [x] What is a "Social Image" of a news article
- [x] How to add a news article
- [x] How to remove a news article
### FAQ
- [x] What to do if the banner item not be shown on home page
- [x] What to do if I can't upload my image: file size too large
## System
- [x] License
- [x] Project status
- [ ] How to perform a backup
- [ ] How to restore a backup
|
1.0
|
User Manual - User Manual related issues:
## Account
- [x] How to add an account
- [x] How to log in
- [x] How to disable an account
- [x] How to change/reset my password
- [ ] How to change/reset other person's password (if you're super administrator)
### FAQ
- [x] What is a good password
- [x] What to do if I have been told that my password has been in a data leak
- [x] What to do if I forgot my password
- [x] What to do if I can't log in
## Catalog
- [x] Good / Bad examples of product / category image
- [x] How to add a category
- [x] How to edit a category
- [x] How to remove a category
- [x] How to add a product
- [x] How to edit a product
- [x] How to remove a product
### FAQ
- [ ] What to do if I can't remove a category: category is not empty
- [x] What to do if I can't upload an image to category or product
- [ ] What to do if the syntax of product description is a mess
- [ ] What to do if I can't upload my image: file size too large
## Campaign
- [x] Good / Bad examples of banner image
- [x] How to add image to home page banner
- [x] How to schedule a banner item
- [x] How to remove an item from home page banner
- [x] Social media best practice
- [x] What is a "Social Image" of a news article
- [x] How to add a news article
- [x] How to remove a news article
### FAQ
- [x] What to do if the banner item not be shown on home page
- [x] What to do if I can't upload my image: file size too large
## System
- [x] License
- [x] Project status
- [ ] How to perform a backup
- [ ] How to restore a backup
|
non_defect
|
user manual user manual related issues account how to add an account how to log in how to disable an account how to change reset my password how to change reset other person s password if you re super administrator faq what is a good password what to do if i have been told that my password has been in a data leak what to do if i forgot my password what to do if i can t log in catalog good bad examples of product category image how to add a category how to edit a category how to remove a category how to add a product how to edit a product how to remove a product faq what to do if i can t remove a category category is not empty what to do if i can t upload an image to category or product what to do if the syntax of product description is a mess what to do if i can t upload my image file size too large campaign good bad examples of banner image how to add image to home page banner how to schedule a banner item how to remove an item from home page banner social media best practice what is a social image of a news article how to add a news article how to remove a news article faq what to do if the banner item not be shown on home page what to do if i can t upload my image file size too large system license project status how to perform a backup how to restore a backup
| 0
|
43,416
| 11,710,025,106
|
IssuesEvent
|
2020-03-08 22:12:13
|
combatopera/aridity
|
https://api.github.com/repos/combatopera/aridity
|
closed
|
new template syntax
|
defect
|
i.e. the rhs of an equals directive, or a template file. current syntax works but i hate it - specifically, having empty string as an alias for get, the layer of indirection for get, and the bespokeness of it all
* consider python template language i.e. {code} with {{ if you want a single curly (and related strict rules)
* consider scheme-like syntax within the curlies, with its various punctuation tricks
* must continue to support barewords, as i don't want users to have to learn quoting rules
* and barewords may contain nested {...} sections
* shell uses barewords so maybe we can steal some of its syntax
|
1.0
|
new template syntax - i.e. the rhs of an equals directive, or a template file. current syntax works but i hate it - specifically, having empty string as an alias for get, the layer of indirection for get, and the bespokeness of it all
* consider python template language i.e. {code} with {{ if you want a single curly (and related strict rules)
* consider scheme-like syntax within the curlies, with its various punctuation tricks
* must continue to support barewords, as i don't want users to have to learn quoting rules
* and barewords may contain nested {...} sections
* shell uses barewords so maybe we can steal some of its syntax
|
defect
|
new template syntax i e the rhs of an equals directive or a template file current syntax works but i hate it specifically having empty string as an alias for get the layer of indirection for get and the bespokeness of it all consider python template language i e code with if you want a single curly and related strict rules consider scheme like syntax within the curlies with its various punctuation tricks must continue to support barewords as i don t want users to have to learn quoting rules and barewords may contain nested sections shell uses barewords so maybe we can steal some of its syntax
| 1
|
68,049
| 21,443,234,442
|
IssuesEvent
|
2022-04-25 01:27:57
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Sending a message while editing a message swallows your sent message
|
T-Defect X-Needs-Design A-Aggregations A-Message-Editing
|
1. Start editing a message, but do not save
2. Send a message using the main composer
3. Goodbye sent message, we missed you
|
1.0
|
Sending a message while editing a message swallows your sent message - 1. Start editing a message, but do not save
2. Send a message using the main composer
3. Goodbye sent message, we missed you
|
defect
|
sending a message while editing a message swallows your sent message start editing a message but do not save send a message using the main composer goodbye sent message we missed you
| 1
|
393,965
| 11,627,350,379
|
IssuesEvent
|
2020-02-27 16:23:36
|
GoogleContainerTools/skaffold
|
https://api.github.com/repos/GoogleContainerTools/skaffold
|
closed
|
Unmark `--status-check` flag as hidden
|
area/ci-cd area/deploy kind/todo priority/p1
|
Related to #3621
One ask/improvement from skaffold survey was to wait for deployments to stabilize.
|
1.0
|
Unmark `--status-check` flag as hidden - Related to #3621
One ask/improvement from skaffold survey was to wait for deployments to stabilize.
|
non_defect
|
unmark status check flag as hidden related to one ask improvement from skaffold survey was to wait for deployments to stabilize
| 0
|
75,774
| 26,041,540,087
|
IssuesEvent
|
2022-12-22 10:50:43
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Element Nightly update failing
|
T-Defect S-Major A-Electron-Nightly O-Uncommon
|
### Steps to reproduce
1. Update the app
https://user-images.githubusercontent.com/769871/209118478-036c435e-d52f-456e-bef2-bd85ce95db62.mov
### Outcome
#### What did you expect?
for the app to be update
#### What happened instead?
It's not updated, and keeps on seeing there's a new version, but it can't download it
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Element Nightly update failing - ### Steps to reproduce
1. Update the app
https://user-images.githubusercontent.com/769871/209118478-036c435e-d52f-456e-bef2-bd85ce95db62.mov
### Outcome
#### What did you expect?
for the app to be update
#### What happened instead?
It's not updated, and keeps on seeing there's a new version, but it can't download it
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
element nightly update failing steps to reproduce update the app outcome what did you expect for the app to be update what happened instead it s not updated and keeps on seeing there s a new version but it can t download it operating system no response application version no response how did you install the app no response homeserver no response will you send logs no
| 1
|
39,732
| 16,082,305,078
|
IssuesEvent
|
2021-04-26 06:59:56
|
Azure/azure-cli
|
https://api.github.com/repos/Azure/azure-cli
|
closed
|
az bicep throws python error on execution
|
ARM Service Attention
|
### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az bicep build`
**Errors:**
```
type object 'datetime.datetime' has no attribute 'fromisoformat'
Traceback (most recent call last):
cli\Lib\site-packages\knack\cli.py, ln 231, in invoke
azure\cli\core\commands\__init__.py, ln 657, in execute
azure\cli\core\commands\__init__.py, ln 720, in _run_jobs_serially
azure\cli\core\commands\__init__.py, ln 691, in _run_job
azure\cli\core\commands\__init__.py, ln 328, in __call__
site-packages\azure\cli\core\__init__.py, ln 807, in default_command_handler
azure\cli\command_modules\resource\custom.py, ln 3294, in build_bicep_file
azure\cli\command_modules\resource\_bicep.py, ln 63, in run_bicep_command
azure\cli\command_modules\resource\_bicep.py, ln 152, in _load_bicep_version_check_result_from_cache
AttributeError: type object 'datetime.datetime' has no attribute 'fromisoformat'
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az bicep build -f {}`
## Expected Behavior
bicep file is compiled to ARM-template.
## Environment Summary
```
Windows-10-10.0.19041-SP0
Python 3.6.8
Installer: MSI
azure-cli 2.22.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
1.0
|
az bicep throws python error on execution -
### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az bicep build`
**Errors:**
```
type object 'datetime.datetime' has no attribute 'fromisoformat'
Traceback (most recent call last):
cli\Lib\site-packages\knack\cli.py, ln 231, in invoke
azure\cli\core\commands\__init__.py, ln 657, in execute
azure\cli\core\commands\__init__.py, ln 720, in _run_jobs_serially
azure\cli\core\commands\__init__.py, ln 691, in _run_job
azure\cli\core\commands\__init__.py, ln 328, in __call__
site-packages\azure\cli\core\__init__.py, ln 807, in default_command_handler
azure\cli\command_modules\resource\custom.py, ln 3294, in build_bicep_file
azure\cli\command_modules\resource\_bicep.py, ln 63, in run_bicep_command
azure\cli\command_modules\resource\_bicep.py, ln 152, in _load_bicep_version_check_result_from_cache
AttributeError: type object 'datetime.datetime' has no attribute 'fromisoformat'
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az bicep build -f {}`
## Expected Behavior
bicep file is compiled to ARM-template.
## Environment Summary
```
Windows-10-10.0.19041-SP0
Python 3.6.8
Installer: MSI
azure-cli 2.22.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
non_defect
|
az bicep throws python error on execution this is autogenerated please review and update as needed describe the bug command name az bicep build errors type object datetime datetime has no attribute fromisoformat traceback most recent call last cli lib site packages knack cli py ln in invoke azure cli core commands init py ln in execute azure cli core commands init py ln in run jobs serially azure cli core commands init py ln in run job azure cli core commands init py ln in call site packages azure cli core init py ln in default command handler azure cli command modules resource custom py ln in build bicep file azure cli command modules resource bicep py ln in run bicep command azure cli command modules resource bicep py ln in load bicep version check result from cache attributeerror type object datetime datetime has no attribute fromisoformat to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az bicep build f expected behavior bicep file is compiled to arm template environment summary windows python installer msi azure cli additional context
| 0
|
114,606
| 24,625,401,675
|
IssuesEvent
|
2022-10-16 13:02:36
|
nmrih/source-game
|
https://api.github.com/repos/nmrih/source-game
|
closed
|
[public-1.12.3] Objective outline takes precedence over inventory outline
|
Status: Reviewed Type: Code Priority: Minimal
|
Aiming at an objective item that's also an inventory item, it's unclear whether you have enough space to pick it up, since the objective outline takes priority (e.g. nmo_anxiety's fire extinguisher). When you're that close to an item, inventory outline should probably take precedence
|
1.0
|
[public-1.12.3] Objective outline takes precedence over inventory outline - Aiming at an objective item that's also an inventory item, it's unclear whether you have enough space to pick it up, since the objective outline takes priority (e.g. nmo_anxiety's fire extinguisher). When you're that close to an item, inventory outline should probably take precedence
|
non_defect
|
objective outline takes precedence over inventory outline aiming at an objective item that s also an inventory item it s unclear whether you have enough space to pick it up since the objective outline takes priority e g nmo anxiety s fire extinguisher when you re that close to an item inventory outline should probably take precedence
| 0
|
49,175
| 13,185,280,704
|
IssuesEvent
|
2020-08-12 21:04:52
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[frame object diff] test fails (Trac #916)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/916
, reported by nega and owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-04-10T02:31:02",
"description": "{{{\n Start 164: frame_object_diff::test_segments.py\n164/329 Test #164: frame_object_diff::test_segments.py ............................***Failed 0.16 sec\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 2: import: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 3: import: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 4: import: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 6: from: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 7: from: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 8: from: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 10: syntax error near unexpected token `('\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 10: `base = os.path.join(os.path.expandvars('$I3_TESTDATA'),'\n}}}",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1428633062869138",
"component": "combo reconstruction",
"summary": "[frame object diff] test fails",
"priority": "normal",
"keywords": "",
"time": "2015-04-10T02:30:41",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[frame object diff] test fails (Trac #916) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/916
, reported by nega and owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-04-10T02:31:02",
"description": "{{{\n Start 164: frame_object_diff::test_segments.py\n164/329 Test #164: frame_object_diff::test_segments.py ............................***Failed 0.16 sec\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 2: import: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 3: import: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 4: import: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 6: from: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 7: from: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 8: from: command not found\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 10: syntax error near unexpected token `('\n/build/buildslave/avnas/quick_icerec_fedora_20/source/frame_object_diff/resources/test/test_segments.py: line 10: `base = os.path.join(os.path.expandvars('$I3_TESTDATA'),'\n}}}",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"_ts": "1428633062869138",
"component": "combo reconstruction",
"summary": "[frame object diff] test fails",
"priority": "normal",
"keywords": "",
"time": "2015-04-10T02:30:41",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
|
defect
|
test fails trac migrated from reported by nega and owned by david schultz json status closed changetime description n start frame object diff test segments py test frame object diff test segments py failed sec n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line import command not found n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line import command not found n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line import command not found n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line from command not found n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line from command not found n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line from command not found n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line syntax error near unexpected token n build buildslave avnas quick icerec fedora source frame object diff resources test test segments py line base os path join os path expandvars testdata n reporter nega cc resolution fixed ts component combo reconstruction summary test fails priority normal keywords time milestone owner david schultz type defect
| 1
|
271,627
| 23,618,673,192
|
IssuesEvent
|
2022-08-24 18:17:03
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
There is a successful activity log when deleting one empty folder in one attached file share which SAS URL is created by an access policy (List, Read)
|
🧪 testing :gear: files
|
**Storage Explorer Version:** 1.26.0-dev
**Build Number:** 20220824.2
**Branch:** main
**Platform/OS:** Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.5 (Apple M1 Pro)
**Architecture:** ia32/x64
**How Found:** Ad-hoc testing
**Regression From:** Not a regression
## Steps to Reproduce ##
1. Expand one storage account -> File Shares.
2. Create a file share -> Create a folder.
3. Right click the file share -> Click 'Manage Access Policies...'.
4. Add one access policy with 'Read, List' permissions -> Generate a SAS URL via the access policy.
5. Attach the file share via the SAS URL.
6. Switch to the attached file share -> Try to delete the folder.
7. Check whether there is a failed activity log.
## Expected Experience ##
There is a failed activity log.
## Actual Experience ##
1. There is a successful activity log
2. The folder still displays.

## Additional Context ##
1. This issue doesn't reproduce when deleting one no-empty folder.
2. This issue doesn't reproduce when attaching via SAS permissions.
|
1.0
|
There is a successful activity log when deleting one empty folder in one attached file share which SAS URL is created by an access policy (List, Read) - **Storage Explorer Version:** 1.26.0-dev
**Build Number:** 20220824.2
**Branch:** main
**Platform/OS:** Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.5 (Apple M1 Pro)
**Architecture:** ia32/x64
**How Found:** Ad-hoc testing
**Regression From:** Not a regression
## Steps to Reproduce ##
1. Expand one storage account -> File Shares.
2. Create a file share -> Create a folder.
3. Right click the file share -> Click 'Manage Access Policies...'.
4. Add one access policy with 'Read, List' permissions -> Generate a SAS URL via the access policy.
5. Attach the file share via the SAS URL.
6. Switch to the attached file share -> Try to delete the folder.
7. Check whether there is a failed activity log.
## Expected Experience ##
There is a failed activity log.
## Actual Experience ##
1. There is a successful activity log
2. The folder still displays.

## Additional Context ##
1. This issue doesn't reproduce when deleting one no-empty folder.
2. This issue doesn't reproduce when attaching via SAS permissions.
|
non_defect
|
there is a successful activity log when deleting one empty folder in one attached file share which sas url is created by an access policy list read storage explorer version dev build number branch main platform os windows linux ubuntu macos monterey apple pro architecture how found ad hoc testing regression from not a regression steps to reproduce expand one storage account file shares create a file share create a folder right click the file share click manage access policies add one access policy with read list permissions generate a sas url via the access policy attach the file share via the sas url switch to the attached file share try to delete the folder check whether there is a failed activity log expected experience there is a failed activity log actual experience there is a successful activity log the folder still displays additional context this issue doesn t reproduce when deleting one no empty folder this issue doesn t reproduce when attaching via sas permissions
| 0
|
33,655
| 7,195,147,580
|
IssuesEvent
|
2018-02-04 14:15:41
|
ganeti/ganeti
|
https://api.github.com/repos/ganeti/ganeti
|
closed
|
ganeti-instance-debootstrap map_disk0 doesn't always map the 1st part
|
AutoMigrated Priority:Medium Status:Fixed Type-Defect
|
Originally reported of Google Code with ID 1124.
```
The map_disk0 function finds the 1st partition by searching for "-1*$blockdev" in kpartx's output.
Problem is the UUID string of other partitions can also have "-1" in them. This bug only manifests when one uses more than one partition on the disk0.
I tripped on this small bug after adding support for 2 partitions on disk0: root FS + swap.
Please find attached a small fix for this.
```
Originally added on 2015-08-26 17:30:16 +0000 UTC.
Attachments:
[ganeti-instance-debootstrap-better-1st-part-match.patch](https://storage.googleapis.com/google-code-attachments/ganeti/issue-1124/comment-0/ganeti-instance-debootstrap-better-1st-part-match.patch)
|
1.0
|
ganeti-instance-debootstrap map_disk0 doesn't always map the 1st part - Originally reported of Google Code with ID 1124.
```
The map_disk0 function finds the 1st partition by searching for "-1*$blockdev" in kpartx's output.
Problem is the UUID string of other partitions can also have "-1" in them. This bug only manifests when one uses more than one partition on the disk0.
I tripped on this small bug after adding support for 2 partitions on disk0: root FS + swap.
Please find attached a small fix for this.
```
Originally added on 2015-08-26 17:30:16 +0000 UTC.
Attachments:
[ganeti-instance-debootstrap-better-1st-part-match.patch](https://storage.googleapis.com/google-code-attachments/ganeti/issue-1124/comment-0/ganeti-instance-debootstrap-better-1st-part-match.patch)
|
defect
|
ganeti instance debootstrap map doesn t always map the part originally reported of google code with id the map function finds the partition by searching for blockdev in kpartx s output problem is the uuid string of other partitions can also have in them this bug only manifests when one uses more than one partition on the i tripped on this small bug after adding support for partitions on root fs swap please find attached a small fix for this originally added on utc attachments
| 1
|
62,584
| 17,082,999,072
|
IssuesEvent
|
2021-07-08 08:16:15
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Showing a room with images in the timeline is very laggy
|
A-Media A-Performance S-Major T-Defect X-Regression
|

On develop.element.io with Firefox 89.0.1
|
1.0
|
Showing a room with images in the timeline is very laggy - 
On develop.element.io with Firefox 89.0.1
|
defect
|
showing a room with images in the timeline is very laggy on develop element io with firefox
| 1
|
19,428
| 3,202,674,075
|
IssuesEvent
|
2015-10-02 15:08:37
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
Hazelcast 3.5.x Memory Leak using IMap
|
Team: Core Type: Critical Type: Defect VERIFIED
|
We are experimenting memory leaks when using IMaps.
Based on our observations, the leak seems related to the fact that we're regularly destroying maps.
Based on Heap analysis, the culprit seems to be around com.hazelcast.map.impl.MapServiceContextImpl.
Setup:
- Hazelcast 3.5.2, JDK 1.7/1.8, reproduced on Linux/Windows 7 and Mac OSX 10.9.5
- 2 nodes with embedded Hazelcast (e.g. not using separate client JVMs)
Scenario:
- One of our node is creating a new IMap roughly once per minute and put stuff in it at regular interval (few hundreds to few thousands per minute).
- Few seconds/minutes later, the other node is removing data from the map, and if the map is empty, destroys the map.
- We do notice a steady growth of the heap usage, which seems related to Hazelcast IMaps internal management.
We did run a couple of variants with the following observations:
- Hazelcast versions 3.5.1 and 3.5.2 exhibits the same memory leak behavior
- We tested both with JMX/ManCenter enabled and disabled, without changes on the results.
- We disable Map "rotation" (e.g. a single map is used instead) and we didn't noticed the memory leak.
- We also tested with version 3.4.6, problem is NOT present - seems something introduced in 3.5.x
We wanted to update from 3.4.x since as our production system is affected by another memory leak documented here: https://github.com/hazelcast/hazelcast/issues/4888
Please find a small project on GitHub which demonstrates the problem: https://github.com/jocelynlepage/hz-map-leak
This project includes a couple of heap dump files (produced with jmap) provided for analysis.
Thanks,
Jocelyn
|
1.0
|
Hazelcast 3.5.x Memory Leak using IMap - We are experimenting memory leaks when using IMaps.
Based on our observations, the leak seems related to the fact that we're regularly destroying maps.
Based on Heap analysis, the culprit seems to be around com.hazelcast.map.impl.MapServiceContextImpl.
Setup:
- Hazelcast 3.5.2, JDK 1.7/1.8, reproduced on Linux/Windows 7 and Mac OSX 10.9.5
- 2 nodes with embedded Hazelcast (e.g. not using separate client JVMs)
Scenario:
- One of our node is creating a new IMap roughly once per minute and put stuff in it at regular interval (few hundreds to few thousands per minute).
- Few seconds/minutes later, the other node is removing data from the map, and if the map is empty, destroys the map.
- We do notice a steady growth of the heap usage, which seems related to Hazelcast IMaps internal management.
We did run a couple of variants with the following observations:
- Hazelcast versions 3.5.1 and 3.5.2 exhibits the same memory leak behavior
- We tested both with JMX/ManCenter enabled and disabled, without changes on the results.
- We disable Map "rotation" (e.g. a single map is used instead) and we didn't noticed the memory leak.
- We also tested with version 3.4.6, problem is NOT present - seems something introduced in 3.5.x
We wanted to update from 3.4.x since as our production system is affected by another memory leak documented here: https://github.com/hazelcast/hazelcast/issues/4888
Please find a small project on GitHub which demonstrates the problem: https://github.com/jocelynlepage/hz-map-leak
This project includes a couple of heap dump files (produced with jmap) provided for analysis.
Thanks,
Jocelyn
|
defect
|
hazelcast x memory leak using imap we are experimenting memory leaks when using imaps based on our observations the leak seems related to the fact that we re regularly destroying maps based on heap analysis the culprit seems to be around com hazelcast map impl mapservicecontextimpl setup hazelcast jdk reproduced on linux windows and mac osx nodes with embedded hazelcast e g not using separate client jvms scenario one of our node is creating a new imap roughly once per minute and put stuff in it at regular interval few hundreds to few thousands per minute few seconds minutes later the other node is removing data from the map and if the map is empty destroys the map we do notice a steady growth of the heap usage which seems related to hazelcast imaps internal management we did run a couple of variants with the following observations hazelcast versions and exhibits the same memory leak behavior we tested both with jmx mancenter enabled and disabled without changes on the results we disable map rotation e g a single map is used instead and we didn t noticed the memory leak we also tested with version problem is not present seems something introduced in x we wanted to update from x since as our production system is affected by another memory leak documented here please find a small project on github which demonstrates the problem this project includes a couple of heap dump files produced with jmap provided for analysis thanks jocelyn
| 1
|
71,492
| 23,652,834,881
|
IssuesEvent
|
2022-08-26 08:25:05
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
closed
|
[FTUE] #2585 makes Chinese users unable to successfully log in or register
|
T-Defect
|
### Steps to reproduce
1. Go to China.
2. Install the latest version of Element.
3. Open Element.
4. ERROR.
### Outcome
#### What did you expect?
I want to use my own homeserver, beacuse matrix.org is blocked in China.
#### What happened instead?
Element will first try to connect to matrix.org, and if it can't, throw an error without first giving me an option to choose my own HS. In fact, I needed to connect to matrix.org via VPN first before Element gave me the option to choose my own HS in the second step. This obviously does not conform to the concept of decentralization.
### Your phone model
moto s20 pro
### Operating system version
a12
### Application version and app store
v1.4.30 and above (F-Droid)
### Homeserver
My own homeserver
### Will you send logs?
No
### Are you willing to provide a PR?
No
|
1.0
|
[FTUE] #2585 makes Chinese users unable to successfully log in or register - ### Steps to reproduce
1. Go to China.
2. Install the latest version of Element.
3. Open Element.
4. ERROR.
### Outcome
#### What did you expect?
I want to use my own homeserver, beacuse matrix.org is blocked in China.
#### What happened instead?
Element will first try to connect to matrix.org, and if it can't, throw an error without first giving me an option to choose my own HS. In fact, I needed to connect to matrix.org via VPN first before Element gave me the option to choose my own HS in the second step. This obviously does not conform to the concept of decentralization.
### Your phone model
moto s20 pro
### Operating system version
a12
### Application version and app store
v1.4.30 and above (F-Droid)
### Homeserver
My own homeserver
### Will you send logs?
No
### Are you willing to provide a PR?
No
|
defect
|
makes chinese users unable to successfully log in or register steps to reproduce go to china install the latest version of element open element error outcome what did you expect i want to use my own homeserver beacuse matrix org is blocked in china what happened instead element will first try to connect to matrix org and if it can t throw an error without first giving me an option to choose my own hs in fact i needed to connect to matrix org via vpn first before element gave me the option to choose my own hs in the second step this obviously does not conform to the concept of decentralization your phone model moto pro operating system version application version and app store and above f droid homeserver my own homeserver will you send logs no are you willing to provide a pr no
| 1
|
79,445
| 3,535,687,595
|
IssuesEvent
|
2016-01-16 18:17:04
|
TheMacLab/game-on
|
https://api.github.com/repos/TheMacLab/game-on
|
opened
|
Giftable messages contain excess spacing above the message body
|
0 - Backlog bug priority-minor
|
Messages sent to users that have received a gift have extra spacing above the message body. The styling is also inconsistent with other messages. They do not turn red when they are unread, nor white when they have been recently read.
<!---
@huboard:{"order":6.467103958129883e-06,"milestone_order":0.060546875}
-->
|
1.0
|
Giftable messages contain excess spacing above the message body - Messages sent to users that have received a gift have extra spacing above the message body. The styling is also inconsistent with other messages. They do not turn red when they are unread, nor white when they have been recently read.
<!---
@huboard:{"order":6.467103958129883e-06,"milestone_order":0.060546875}
-->
|
non_defect
|
giftable messages contain excess spacing above the message body messages sent to users that have received a gift have extra spacing above the message body the styling is also inconsistent with other messages they do not turn red when they are unread nor white when they have been recently read huboard order milestone order
| 0
|
2,425
| 2,607,902,121
|
IssuesEvent
|
2015-02-26 00:14:02
|
chrsmithdemos/zen-coding
|
https://api.github.com/repos/chrsmithdemos/zen-coding
|
closed
|
Data:URL actions
|
auto-migrated Milestone-0.7 Priority-Medium Type-Defect
|
```
I can think about two actions for work with Data:URI with zen-
coding:
1. Convert image to Data:URL from address in <img/> and
background url().
2. Preview Data:URI (or any images?) from these addresses.
Although like updating <img> size this action maybe won't be
available on every editor.
```
-----
Original issue reported on code.google.com by `kizmarh` on 25 Feb 2010 at 4:11
|
1.0
|
Data:URL actions - ```
I can think about two actions for work with Data:URI with zen-
coding:
1. Convert image to Data:URL from address in <img/> and
background url().
2. Preview Data:URI (or any images?) from these addresses.
Although like updating <img> size this action maybe won't be
available on every editor.
```
-----
Original issue reported on code.google.com by `kizmarh` on 25 Feb 2010 at 4:11
|
defect
|
data url actions i can think about two actions for work with data uri with zen coding convert image to data url from address in and background url preview data uri or any images from these addresses although like updating size this action maybe won t be available on every editor original issue reported on code google com by kizmarh on feb at
| 1
|
33,528
| 7,153,791,335
|
IssuesEvent
|
2018-01-26 04:03:51
|
line/armeria
|
https://api.github.com/repos/line/armeria
|
closed
|
ClientFactory.close doesn't actually close in-progress connections
|
defect
|
I was trying to simulate a scenario where a client opens a connection to the server (e.g., starts a gRPC stream) but then forcibly closes the socket. I would expect `ClientFactory.close` to close the socket, but it doesn't actually seem to. It calls `DefaultKeyedChannelPool.close`, but that only keeps track of available connections, not offered connections, so the actual stream doesn't seem to be affected. Should the pool also keep track of offered connections to allow closing them when needed?
|
1.0
|
ClientFactory.close doesn't actually close in-progress connections - I was trying to simulate a scenario where a client opens a connection to the server (e.g., starts a gRPC stream) but then forcibly closes the socket. I would expect `ClientFactory.close` to close the socket, but it doesn't actually seem to. It calls `DefaultKeyedChannelPool.close`, but that only keeps track of available connections, not offered connections, so the actual stream doesn't seem to be affected. Should the pool also keep track of offered connections to allow closing them when needed?
|
defect
|
clientfactory close doesn t actually close in progress connections i was trying to simulate a scenario where a client opens a connection to the server e g starts a grpc stream but then forcibly closes the socket i would expect clientfactory close to close the socket but it doesn t actually seem to it calls defaultkeyedchannelpool close but that only keeps track of available connections not offered connections so the actual stream doesn t seem to be affected should the pool also keep track of offered connections to allow closing them when needed
| 1
|
68,098
| 21,477,564,213
|
IssuesEvent
|
2022-04-26 14:49:50
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Timeline jumps when switching rooms to timelines with potential URL previews
|
T-Defect
|
### Steps to reproduce
1. Disable URL previews server-side, but not in the client
2. Visit a room with lots of links (in my case, EMS Ops)
3. Note the timeline jump
### Outcome
#### What did you expect?
For the timeline not to jump immediately upon opening the room.
#### What happened instead?
The timeline jumped. It appears to be a gap roughly the height of the typing notification container, and goes away after the URL previews error out.
This appears to be a regression: we didn't previously reserve the space while the preview was loading, instead opting to pop-in the previews.
### Operating system
Windows 10
### Application version
0.0.1-nightly.2022042601
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No
|
1.0
|
Timeline jumps when switching rooms to timelines with potential URL previews - ### Steps to reproduce
1. Disable URL previews server-side, but not in the client
2. Visit a room with lots of links (in my case, EMS Ops)
3. Note the timeline jump
### Outcome
#### What did you expect?
For the timeline not to jump immediately upon opening the room.
#### What happened instead?
The timeline jumped. It appears to be a gap roughly the height of the typing notification container, and goes away after the URL previews error out.
This appears to be a regression: we didn't previously reserve the space while the preview was loading, instead opting to pop-in the previews.
### Operating system
Windows 10
### Application version
0.0.1-nightly.2022042601
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No
|
defect
|
timeline jumps when switching rooms to timelines with potential url previews steps to reproduce disable url previews server side but not in the client visit a room with lots of links in my case ems ops note the timeline jump outcome what did you expect for the timeline not to jump immediately upon opening the room what happened instead the timeline jumped it appears to be a gap roughly the height of the typing notification container and goes away after the url previews error out this appears to be a regression we didn t previously reserve the space while the preview was loading instead opting to pop in the previews operating system windows application version nightly how did you install the app the internet homeserver io will you send logs no
| 1
|
46,995
| 13,056,011,500
|
IssuesEvent
|
2020-07-30 03:23:10
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
Subtracting bias in photospline/I3SplineTable even if derivatives are available (Trac #2236)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
Migrated from https://code.icecube.wisc.edu/ticket/2236
```json
{
"status": "closed",
"changetime": "2019-02-04T17:52:51",
"description": "After evaluating photosplines a bias must only be subtracted if no derivatives are available but it is always done.\nI am going to commit a fix for that soon.",
"reporter": "mhieronymus",
"cc": "",
"resolution": "fixed",
"_ts": "1549302771596052",
"component": "combo reconstruction",
"summary": "Subtracting bias in photospline/I3SplineTable even if derivatives are available",
"priority": "normal",
"keywords": "Bug I3SplineTable",
"time": "2019-01-31T08:17:04",
"milestone": "",
"owner": "mhieronymus",
"type": "defect"
}
```
|
1.0
|
Subtracting bias in photospline/I3SplineTable even if derivatives are available (Trac #2236) - Migrated from https://code.icecube.wisc.edu/ticket/2236
```json
{
"status": "closed",
"changetime": "2019-02-04T17:52:51",
"description": "After evaluating photosplines a bias must only be subtracted if no derivatives are available but it is always done.\nI am going to commit a fix for that soon.",
"reporter": "mhieronymus",
"cc": "",
"resolution": "fixed",
"_ts": "1549302771596052",
"component": "combo reconstruction",
"summary": "Subtracting bias in photospline/I3SplineTable even if derivatives are available",
"priority": "normal",
"keywords": "Bug I3SplineTable",
"time": "2019-01-31T08:17:04",
"milestone": "",
"owner": "mhieronymus",
"type": "defect"
}
```
|
defect
|
subtracting bias in photospline even if derivatives are available trac migrated from json status closed changetime description after evaluating photosplines a bias must only be subtracted if no derivatives are available but it is always done ni am going to commit a fix for that soon reporter mhieronymus cc resolution fixed ts component combo reconstruction summary subtracting bias in photospline even if derivatives are available priority normal keywords bug time milestone owner mhieronymus type defect
| 1
|
70,339
| 23,129,550,246
|
IssuesEvent
|
2022-07-28 09:08:42
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
DataTable: resetting FilterMetal#filterValue to null wont be applied to filter component
|
:lady_beetle: defect
|
### Describe the bug
the old value is still displayed but the filter applied correctly
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
_No response_
### PrimeFaces version
12.0.0-RC2
### Theme
_No response_
### JSF implementation
_No response_
### JSF version
_No response_
### Browser(s)
_No response_
|
1.0
|
DataTable: resetting FilterMetal#filterValue to null wont be applied to filter component - ### Describe the bug
the old value is still displayed but the filter applied correctly
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
_No response_
### PrimeFaces version
12.0.0-RC2
### Theme
_No response_
### JSF implementation
_No response_
### JSF version
_No response_
### Browser(s)
_No response_
|
defect
|
datatable resetting filtermetal filtervalue to null wont be applied to filter component describe the bug the old value is still displayed but the filter applied correctly reproducer no response expected behavior no response primefaces edition no response primefaces version theme no response jsf implementation no response jsf version no response browser s no response
| 1
|
361,789
| 25,351,273,170
|
IssuesEvent
|
2022-11-19 20:07:26
|
jorgenriseth/multicompartment-solute-transport
|
https://api.github.com/repos/jorgenriseth/multicompartment-solute-transport
|
opened
|
Add license
|
documentation
|
The code should have a licence which clarifies how people should credit the code if they use/adapt it.
@vegarvi any thoughts?
See for instance: https://choosealicense.com/
|
1.0
|
Add license - The code should have a licence which clarifies how people should credit the code if they use/adapt it.
@vegarvi any thoughts?
See for instance: https://choosealicense.com/
|
non_defect
|
add license the code should have a licence which clarifies how people should credit the code if they use adapt it vegarvi any thoughts see for instance
| 0
|
10,903
| 2,622,846,097
|
IssuesEvent
|
2015-03-04 08:02:40
|
max99x/pagemon-chrome-ext
|
https://api.github.com/repos/max99x/pagemon-chrome-ext
|
closed
|
Option amnesia
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
Not quite sure. I changed the sorting option to be by last changed, the check
interval to 1 day and the view all function to "view changes". After a reboot,
possibly when Chrome was not shut down gracefully (not sure), options were
partially forgotten.
In particular, the view all function reverted to "view originals" and the check
interval reverted to 2 hours and 59 mins (I think).
What is the expected output? What do you see instead?
Settings should be saved either immediately (or there should be a "save"
button".
What version of the Chrome are you using? On what operating system?
6.0.453.1 with Windows XP Pro SP3
Please provide any additional information below.
```
Original issue reported on code.google.com by `gkaemp...@gmail.com` on 4 Jul 2010 at 2:25
|
1.0
|
Option amnesia - ```
What steps will reproduce the problem?
Not quite sure. I changed the sorting option to be by last changed, the check
interval to 1 day and the view all function to "view changes". After a reboot,
possibly when Chrome was not shut down gracefully (not sure), options were
partially forgotten.
In particular, the view all function reverted to "view originals" and the check
interval reverted to 2 hours and 59 mins (I think).
What is the expected output? What do you see instead?
Settings should be saved either immediately (or there should be a "save"
button".
What version of the Chrome are you using? On what operating system?
6.0.453.1 with Windows XP Pro SP3
Please provide any additional information below.
```
Original issue reported on code.google.com by `gkaemp...@gmail.com` on 4 Jul 2010 at 2:25
|
defect
|
option amnesia what steps will reproduce the problem not quite sure i changed the sorting option to be by last changed the check interval to day and the view all function to view changes after a reboot possibly when chrome was not shut down gracefully not sure options were partially forgotten in particular the view all function reverted to view originals and the check interval reverted to hours and mins i think what is the expected output what do you see instead settings should be saved either immediately or there should be a save button what version of the chrome are you using on what operating system with windows xp pro please provide any additional information below original issue reported on code google com by gkaemp gmail com on jul at
| 1
|
54,368
| 13,626,793,102
|
IssuesEvent
|
2020-09-24 11:37:06
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
Overly long TXT record breaks lmdbbackend import
|
auth defect
|
- Program: Authoritative
- Issue type: Bug report
### Short description
Long TXT records cause lmdb import process to break with the error message:
```
$ pdnsutil load-zone example.com example.com.db
Error: putting data: MDB_BAD_VALSIZE: Unsupported size of key/DB name/data, or wrong DUPFIXED size
```
Example zone file `example.com.db`:
```
example.com. 86400 IN SOA example.com. example.com. (
2018031306 ;Serial Number
14400 ;refresh
7200 ;retry
3600000 ;expire
14400 )
example.com. 10 IN TXT "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa"
```
I guess you are storing the TXT aspect of the detail in the lmdb key record which causes it to go over the 512 byte key limit? If this is the case it may need a db design change before launching the new lmdb backend as I suspect TXT records should be stored as values.
### Environment
- Operating system: centos7
- Software version: pdns-4.2.0-rc2.169.master.g85474361f-centos-7.tar.bz2
|
1.0
|
Overly long TXT record breaks lmdbbackend import - - Program: Authoritative
- Issue type: Bug report
### Short description
Long TXT records cause lmdb import process to break with the error message:
```
$ pdnsutil load-zone example.com example.com.db
Error: putting data: MDB_BAD_VALSIZE: Unsupported size of key/DB name/data, or wrong DUPFIXED size
```
Example zone file `example.com.db`:
```
example.com. 86400 IN SOA example.com. example.com. (
2018031306 ;Serial Number
14400 ;refresh
7200 ;retry
3600000 ;expire
14400 )
example.com. 10 IN TXT "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa"
```
I guess you are storing the TXT aspect of the detail in the lmdb key record which causes it to go over the 512 byte key limit? If this is the case it may need a db design change before launching the new lmdb backend as I suspect TXT records should be stored as values.
### Environment
- Operating system: centos7
- Software version: pdns-4.2.0-rc2.169.master.g85474361f-centos-7.tar.bz2
|
defect
|
overly long txt record breaks lmdbbackend import program authoritative issue type bug report short description long txt records cause lmdb import process to break with the error message pdnsutil load zone example com example com db error putting data mdb bad valsize unsupported size of key db name data or wrong dupfixed size example zone file example com db example com in soa example com example com serial number refresh retry expire example com in txt aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa i guess you are storing the txt aspect of the detail in the lmdb key record which causes it to go over the byte key limit if this is the case it may need a db design change before launching the new lmdb backend as i suspect txt records should be stored as values environment operating system software version pdns master centos tar
| 1
|
21,754
| 14,786,431,571
|
IssuesEvent
|
2021-01-12 05:29:43
|
pol-is/polis
|
https://api.github.com/repos/pol-is/polis
|
opened
|
Use GitHub Container Registry to store pre-built containers for test workflows
|
⚒️ infrastructure
|
Right now, we're pushing nightly builds to docker hub. In theory, this makes deploying quicker for new people, since can pull instead of building. In practice, I don't think it's used much.
Having said that, we're also building containers in order to run cypress tests. If we start building them for cross-browser testing of old browsers on BrowserStack, then we'll be building the containers at least twice per commit. This expends twice as many build minutes as we need, and each docker build takes about 8 minutes.
We could instead build the containers in one workflow, and push them to GitHub Container Registry. These would be set to private, so just for internal tests. We could then pull them in the workflows that needs to spin up an instance, without rebuilding them each time (e.g. cypress tests, browserstack tests, etc).
GitHub Container Registry vs GitHub Docker Registry: https://docs.github.com/en/free-pro-team@latest/packages/guides/migrating-to-github-container-registry-for-docker-images (main thing is that GHCR has more fine-grained permissions)
Enabling: https://docs.github.com/en/free-pro-team@latest/packages/guides/enabling-improved-container-support
|
1.0
|
Use GitHub Container Registry to store pre-built containers for test workflows - Right now, we're pushing nightly builds to docker hub. In theory, this makes deploying quicker for new people, since can pull instead of building. In practice, I don't think it's used much.
Having said that, we're also building containers in order to run cypress tests. If we start building them for cross-browser testing of old browsers on BrowserStack, then we'll be building the containers at least twice per commit. This expends twice as many build minutes as we need, and each docker build takes about 8 minutes.
We could instead build the containers in one workflow, and push them to GitHub Container Registry. These would be set to private, so just for internal tests. We could then pull them in the workflows that needs to spin up an instance, without rebuilding them each time (e.g. cypress tests, browserstack tests, etc).
GitHub Container Registry vs GitHub Docker Registry: https://docs.github.com/en/free-pro-team@latest/packages/guides/migrating-to-github-container-registry-for-docker-images (main thing is that GHCR has more fine-grained permissions)
Enabling: https://docs.github.com/en/free-pro-team@latest/packages/guides/enabling-improved-container-support
|
non_defect
|
use github container registry to store pre built containers for test workflows right now we re pushing nightly builds to docker hub in theory this makes deploying quicker for new people since can pull instead of building in practice i don t think it s used much having said that we re also building containers in order to run cypress tests if we start building them for cross browser testing of old browsers on browserstack then we ll be building the containers at least twice per commit this expends twice as many build minutes as we need and each docker build takes about minutes we could instead build the containers in one workflow and push them to github container registry these would be set to private so just for internal tests we could then pull them in the workflows that needs to spin up an instance without rebuilding them each time e g cypress tests browserstack tests etc github container registry vs github docker registry main thing is that ghcr has more fine grained permissions enabling
| 0
|
65,234
| 8,793,749,296
|
IssuesEvent
|
2018-12-21 21:19:42
|
aws/aws-sdk-js
|
https://api.github.com/repos/aws/aws-sdk-js
|
closed
|
Aws-sdk incompatibility with google-closure-compiler
|
documentation in-progress third-party
|
There are 2 places where aws-sdk uses @abstract annotation:
(1) https://github.com/aws/aws-sdk-js/blob/2a872adac4f1c49d45feb5b368e72adf0c343cf4/lib/service.js#L11
(2) https://github.com/aws/aws-sdk-js/blob/2a872adac4f1c49d45feb5b368e72adf0c343cf4/lib/service.js#L431
Google closure compiler has a custom meaning for methods marked with @abstract attribute:
https://github.com/google/closure-compiler/wiki/@abstract-classes-and-methods
Currently attempt to use google-closure-compiler on aws-sdk results in stripping code that annotated with @abstract. Because of that and (1), aws-sdk stops working
Please address the issue of this incompatibility, since it is easy to remove these 2 lines from aws-sdk
|
1.0
|
Aws-sdk incompatibility with google-closure-compiler - There are 2 places where aws-sdk uses @abstract annotation:
(1) https://github.com/aws/aws-sdk-js/blob/2a872adac4f1c49d45feb5b368e72adf0c343cf4/lib/service.js#L11
(2) https://github.com/aws/aws-sdk-js/blob/2a872adac4f1c49d45feb5b368e72adf0c343cf4/lib/service.js#L431
Google closure compiler has a custom meaning for methods marked with @abstract attribute:
https://github.com/google/closure-compiler/wiki/@abstract-classes-and-methods
Currently attempt to use google-closure-compiler on aws-sdk results in stripping code that annotated with @abstract. Because of that and (1), aws-sdk stops working
Please address the issue of this incompatibility, since it is easy to remove these 2 lines from aws-sdk
|
non_defect
|
aws sdk incompatibility with google closure compiler there are places where aws sdk uses abstract annotation google closure compiler has a custom meaning for methods marked with abstract attribute currently attempt to use google closure compiler on aws sdk results in stripping code that annotated with abstract because of that and aws sdk stops working please address the issue of this incompatibility since it is easy to remove these lines from aws sdk
| 0
|
831,319
| 32,045,074,292
|
IssuesEvent
|
2023-09-23 00:16:45
|
grousewood-games/solforge
|
https://api.github.com/repos/grousewood-games/solforge
|
closed
|
Binben Wyrms don't self destruct
|
needs release :butterfly: bug :bug: priority :pizza:
|
Recent refactor have the worms attempting to handle their destruction instead of Binben. Nothing happens so they remain alive.
|
1.0
|
Binben Wyrms don't self destruct - Recent refactor have the worms attempting to handle their destruction instead of Binben. Nothing happens so they remain alive.
|
non_defect
|
binben wyrms don t self destruct recent refactor have the worms attempting to handle their destruction instead of binben nothing happens so they remain alive
| 0
|
25,790
| 4,446,060,057
|
IssuesEvent
|
2016-08-20 12:36:31
|
jliljebl/flowblade
|
https://api.github.com/repos/jliljebl/flowblade
|
reopened
|
Segfault - filter "freeze" + pre-render
|
auto-migrated bug Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Apply the filter "freeze" ("geler" in french) to a sequence
2. Place your cursor on this sequence to pre-render it
What is the expected output? What do you see instead?
What is expected : A pre-render of the sequence with the filter.
What happen : a segmentation fault
What version of the product are you using? On what operating system?
Flowblade 0.12.0
Linux Mint 15, MATE 1.6.0
Please provide any additional information below.
This also happen when the cursor is already on the sequence and you use this
filter.
I didn't manage to use gdb for further informations, because /usr/bin/flowblade
is a shell script.
(also, I know this isn't the right place, but... Where can I find help on
flowblade. I want to do a specific task and I don't kow how...)
```
Original issue reported on code.google.com by `tapata.t...@gmail.com` on 29 Jun 2014 at 3:38
|
1.0
|
Segfault - filter "freeze" + pre-render - ```
What steps will reproduce the problem?
1. Apply the filter "freeze" ("geler" in french) to a sequence
2. Place your cursor on this sequence to pre-render it
What is the expected output? What do you see instead?
What is expected : A pre-render of the sequence with the filter.
What happen : a segmentation fault
What version of the product are you using? On what operating system?
Flowblade 0.12.0
Linux Mint 15, MATE 1.6.0
Please provide any additional information below.
This also happen when the cursor is already on the sequence and you use this
filter.
I didn't manage to use gdb for further informations, because /usr/bin/flowblade
is a shell script.
(also, I know this isn't the right place, but... Where can I find help on
flowblade. I want to do a specific task and I don't kow how...)
```
Original issue reported on code.google.com by `tapata.t...@gmail.com` on 29 Jun 2014 at 3:38
|
defect
|
segfault filter freeze pre render what steps will reproduce the problem apply the filter freeze geler in french to a sequence place your cursor on this sequence to pre render it what is the expected output what do you see instead what is expected a pre render of the sequence with the filter what happen a segmentation fault what version of the product are you using on what operating system flowblade linux mint mate please provide any additional information below this also happen when the cursor is already on the sequence and you use this filter i didn t manage to use gdb for further informations because usr bin flowblade is a shell script also i know this isn t the right place but where can i find help on flowblade i want to do a specific task and i don t kow how original issue reported on code google com by tapata t gmail com on jun at
| 1
|
62,774
| 17,193,332,081
|
IssuesEvent
|
2021-07-16 14:02:06
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
Incident beam radiation calculation in the green roof model
|
Defect
|
Issue overview
--------------
In the current green roof (eco-roof) model, the solar incidence angle was not considered for the calculation of the shortwave radiation, specifically the incident beam radiation term. This shortwave radiation was used in both vegetation and soil energy balance equations in this model for green roofs. This could affect plant evapotranspiration and temperatures for soil surface and foliage predicted from the model.
### Details
Some additional details for this issue (if relevant):
- v9.5
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
1.0
|
Incident beam radiation calculation in the green roof model - Issue overview
--------------
In the current green roof (eco-roof) model, the solar incidence angle was not considered for the calculation of the shortwave radiation, specifically the incident beam radiation term. This shortwave radiation was used in both vegetation and soil energy balance equations in this model for green roofs. This could affect plant evapotranspiration and temperatures for soil surface and foliage predicted from the model.
### Details
Some additional details for this issue (if relevant):
- v9.5
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
defect
|
incident beam radiation calculation in the green roof model issue overview in the current green roof eco roof model the solar incidence angle was not considered for the calculation of the shortwave radiation specifically the incident beam radiation term this shortwave radiation was used in both vegetation and soil energy balance equations in this model for green roofs this could affect plant evapotranspiration and temperatures for soil surface and foliage predicted from the model details some additional details for this issue if relevant checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
| 1
|
5,499
| 2,610,188,760
|
IssuesEvent
|
2015-02-26 18:59:48
|
chrsmith/quchuseban
|
https://api.github.com/repos/chrsmith/quchuseban
|
opened
|
介绍色斑怎样可以消除
|
auto-migrated Priority-Medium Type-Defect
|
```
《摘要》
白皙的皮肤上假如多了些星星点点的雀斑,会让你的魅力值��
�减。而且用任何化妆品也是掩盖不了的。就算可以也只是一�
��的。我们能通过医疗美容的方法来治疗,中药祛斑最近颇受
欢迎,为广大雀斑女性解决了肌肤问题,中药祛斑产品哪种��
�好呢?色斑怎样可以消除,
《客户案例》
杭女士 28岁<br>
我一直以为自己想要去斑根本是痴人说梦,试了那么多��
�方法,用了那么多的东西,还是没有撼动脸上的斑分毫,直�
��使用了「黛芙薇尔精华液」,带给我从来没有过的淡斑效果
,才让我觉得去斑真的有望。下面是我使用「黛芙薇尔精华��
�」祛斑的经历。<br>
本来我脸上的斑从小就有,医生说遗传斑是血液的问题��
�去斑的希望很渺茫,我也尝试了很多方法,真的跟医生说的�
��样,希望渺茫。去年的时候脸上的斑不仅越来越大,而且越
来越多,原来的时候只是鼻梁周围有点,后来是脸额颊上也��
�了,我也不知道是什么原因形成的,总之每次照镜子都会看�
��脸上的斑,心里很着急。病急乱投医,到处寻找祛斑的方法
无果,我只好求助于网络,通过搜索知道了「黛芙薇尔精华��
�」,看了「黛芙薇尔精华液」的产品介绍和说明,感觉很可�
��,而且蛮符合我的情况,就在网上定购了一个周期的看看效
果。<br>
网上对「黛芙薇尔精华液」的功效评价都是说效果不错��
�就是时间慢了点,要一直坚持。其实只要能去掉脸上的斑,�
��是使用产品三四年我也愿意。前几天使用的时候真的感觉不
到很大的效果,就是皮肤好像比较有光泽了,没有那么暗沉��
�专家的说法是祛斑套装里配了排毒精华液,会排出体内毒素�
��这样一来就能从根源改善皮肤的健康状况了。使用祛斑套装
一个多星期以后,感觉固守在脸上的斑有变淡的趋势,深褐��
�的一块东西变成了褐色,颜色没那么深,看到这情况真的很�
��心。我继续坚持使用套装,一个月以后,脸上的斑真的淡化
得比较明显了,连身边的人都问我用了什么产品,淡斑效果��
�么好。
使用完第一个周期之后,自我感觉还是相当不错的,我就定��
�了第二个周期的套装继续使用,等我把两个周期的套装都使�
��完之后,脸上的斑变得很淡很淡,而且我的皮肤还变好了呢
,以前白皙的皮肤现在变得更白,由于没有斑斑点点影响容��
�,我整个人焕然一新,变得非常有自信。医生说斑只能等到�
��化后才能去掉,使用两个周期的「黛芙薇尔精华液」祛斑套
装已经让我成功淡斑,为了彻底祛斑我继续使用第三个周期��
�套装,在第三个周期的祛斑套装使用过程中,真的有祛斑的�
��效,那些原本淡化的斑慢慢消失不见了,好像一夜之间被皮
肤抚平了一样,真的很开心。据自己的祛斑经历,个人觉得��
�虽然祛斑产品那么多,但还是「黛芙薇尔精华液」效果最好�
��
阅读了色斑怎样可以消除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎样可以消除,同时为您分享祛斑小方法
要想斑点不形成,那么就要调整好身体的机能,均衡饮食,��
�持足够的睡眠,要多喝水、多吃水果、适当补充糖分、多吃�
��质蛋白质的食物,对皮肤的光滑细腻很有帮助
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:33
|
1.0
|
介绍色斑怎样可以消除 - ```
《摘要》
白皙的皮肤上假如多了些星星点点的雀斑,会让你的魅力值��
�减。而且用任何化妆品也是掩盖不了的。就算可以也只是一�
��的。我们能通过医疗美容的方法来治疗,中药祛斑最近颇受
欢迎,为广大雀斑女性解决了肌肤问题,中药祛斑产品哪种��
�好呢?色斑怎样可以消除,
《客户案例》
杭女士 28岁<br>
我一直以为自己想要去斑根本是痴人说梦,试了那么多��
�方法,用了那么多的东西,还是没有撼动脸上的斑分毫,直�
��使用了「黛芙薇尔精华液」,带给我从来没有过的淡斑效果
,才让我觉得去斑真的有望。下面是我使用「黛芙薇尔精华��
�」祛斑的经历。<br>
本来我脸上的斑从小就有,医生说遗传斑是血液的问题��
�去斑的希望很渺茫,我也尝试了很多方法,真的跟医生说的�
��样,希望渺茫。去年的时候脸上的斑不仅越来越大,而且越
来越多,原来的时候只是鼻梁周围有点,后来是脸额颊上也��
�了,我也不知道是什么原因形成的,总之每次照镜子都会看�
��脸上的斑,心里很着急。病急乱投医,到处寻找祛斑的方法
无果,我只好求助于网络,通过搜索知道了「黛芙薇尔精华��
�」,看了「黛芙薇尔精华液」的产品介绍和说明,感觉很可�
��,而且蛮符合我的情况,就在网上定购了一个周期的看看效
果。<br>
网上对「黛芙薇尔精华液」的功效评价都是说效果不错��
�就是时间慢了点,要一直坚持。其实只要能去掉脸上的斑,�
��是使用产品三四年我也愿意。前几天使用的时候真的感觉不
到很大的效果,就是皮肤好像比较有光泽了,没有那么暗沉��
�专家的说法是祛斑套装里配了排毒精华液,会排出体内毒素�
��这样一来就能从根源改善皮肤的健康状况了。使用祛斑套装
一个多星期以后,感觉固守在脸上的斑有变淡的趋势,深褐��
�的一块东西变成了褐色,颜色没那么深,看到这情况真的很�
��心。我继续坚持使用套装,一个月以后,脸上的斑真的淡化
得比较明显了,连身边的人都问我用了什么产品,淡斑效果��
�么好。
使用完第一个周期之后,自我感觉还是相当不错的,我就定��
�了第二个周期的套装继续使用,等我把两个周期的套装都使�
��完之后,脸上的斑变得很淡很淡,而且我的皮肤还变好了呢
,以前白皙的皮肤现在变得更白,由于没有斑斑点点影响容��
�,我整个人焕然一新,变得非常有自信。医生说斑只能等到�
��化后才能去掉,使用两个周期的「黛芙薇尔精华液」祛斑套
装已经让我成功淡斑,为了彻底祛斑我继续使用第三个周期��
�套装,在第三个周期的祛斑套装使用过程中,真的有祛斑的�
��效,那些原本淡化的斑慢慢消失不见了,好像一夜之间被皮
肤抚平了一样,真的很开心。据自己的祛斑经历,个人觉得��
�虽然祛斑产品那么多,但还是「黛芙薇尔精华液」效果最好�
��
阅读了色斑怎样可以消除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎样可以消除,同时为您分享祛斑小方法
要想斑点不形成,那么就要调整好身体的机能,均衡饮食,��
�持足够的睡眠,要多喝水、多吃水果、适当补充糖分、多吃�
��质蛋白质的食物,对皮肤的光滑细腻很有帮助
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:33
|
defect
|
介绍色斑怎样可以消除 《摘要》 白皙的皮肤上假如多了些星星点点的雀斑,会让你的魅力值�� �减。而且用任何化妆品也是掩盖不了的。就算可以也只是一� ��的。我们能通过医疗美容的方法来治疗,中药祛斑最近颇受 欢迎,为广大雀斑女性解决了肌肤问题,中药祛斑产品哪种�� �好呢 色斑怎样可以消除, 《客户案例》 杭女士 我一直以为自己想要去斑根本是痴人说梦,试了那么多�� �方法,用了那么多的东西,还是没有撼动脸上的斑分毫,直� ��使用了「黛芙薇尔精华液」,带给我从来没有过的淡斑效果 ,才让我觉得去斑真的有望。下面是我使用「黛芙薇尔精华�� �」祛斑的经历。 本来我脸上的斑从小就有,医生说遗传斑是血液的问题�� �去斑的希望很渺茫,我也尝试了很多方法,真的跟医生说的� ��样,希望渺茫。去年的时候脸上的斑不仅越来越大,而且越 来越多,原来的时候只是鼻梁周围有点,后来是脸额颊上也�� �了,我也不知道是什么原因形成的,总之每次照镜子都会看� ��脸上的斑,心里很着急。病急乱投医,到处寻找祛斑的方法 无果,我只好求助于网络,通过搜索知道了「黛芙薇尔精华�� �」,看了「黛芙薇尔精华液」的产品介绍和说明,感觉很可� ��,而且蛮符合我的情况,就在网上定购了一个周期的看看效 果。 网上对「黛芙薇尔精华液」的功效评价都是说效果不错�� �就是时间慢了点,要一直坚持。其实只要能去掉脸上的斑,� ��是使用产品三四年我也愿意。前几天使用的时候真的感觉不 到很大的效果,就是皮肤好像比较有光泽了,没有那么暗沉�� �专家的说法是祛斑套装里配了排毒精华液,会排出体内毒素� ��这样一来就能从根源改善皮肤的健康状况了。使用祛斑套装 一个多星期以后,感觉固守在脸上的斑有变淡的趋势,深褐�� �的一块东西变成了褐色,颜色没那么深,看到这情况真的很� ��心。我继续坚持使用套装,一个月以后,脸上的斑真的淡化 得比较明显了,连身边的人都问我用了什么产品,淡斑效果�� �么好。 使用完第一个周期之后,自我感觉还是相当不错的,我就定�� �了第二个周期的套装继续使用,等我把两个周期的套装都使� ��完之后,脸上的斑变得很淡很淡,而且我的皮肤还变好了呢 ,以前白皙的皮肤现在变得更白,由于没有斑斑点点影响容�� �,我整个人焕然一新,变得非常有自信。医生说斑只能等到� ��化后才能去掉,使用两个周期的「黛芙薇尔精华液」祛斑套 装已经让我成功淡斑,为了彻底祛斑我继续使用第三个周期�� �套装,在第三个周期的祛斑套装使用过程中,真的有祛斑的� ��效,那些原本淡化的斑慢慢消失不见了,好像一夜之间被皮 肤抚平了一样,真的很开心。据自己的祛斑经历,个人觉得�� �虽然祛斑产品那么多,但还是「黛芙薇尔精华液」效果最好� �� 阅读了色斑怎样可以消除,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 色斑怎样可以消除,同时为您分享祛斑小方法 要想斑点不形成,那么就要调整好身体的机能,均衡饮食,�� �持足够的睡眠,要多喝水、多吃水果、适当补充糖分、多吃� ��质蛋白质的食物,对皮肤的光滑细腻很有帮助 original issue reported on code google com by additive gmail com on jul at
| 1
|
75,171
| 25,567,867,288
|
IssuesEvent
|
2022-11-30 15:30:48
|
vector-im/element-ios
|
https://api.github.com/repos/vector-im/element-ios
|
closed
|
Rich Text Composer: Sometimes Fullscreen mode breaks when opening it when keyboard is not showing
|
T-Defect S-Minor O-Occasional Z-Labs A-Rich-Text-Editor
|
### Steps to reproduce
1. Open any chat with RTE enabled
2. Tap on fullscreen toggle when keyboard is not shown (either dismiss it first or never bring it up)
### Outcome
#### What did you expect?
Full screen mode shows correctly
#### What happened instead?
Sometimes it shows in a broken state, here is an example:

It's still possible to dismiss this view by dragging it to the bottom.
### Your phone model
iPhone 13 Pro Max
### Operating system version
iOS 16.1
### Application version
develop
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Rich Text Composer: Sometimes Fullscreen mode breaks when opening it when keyboard is not showing - ### Steps to reproduce
1. Open any chat with RTE enabled
2. Tap on fullscreen toggle when keyboard is not shown (either dismiss it first or never bring it up)
### Outcome
#### What did you expect?
Full screen mode shows correctly
#### What happened instead?
Sometimes it shows in a broken state, here is an example:

It's still possible to dismiss this view by dragging it to the bottom.
### Your phone model
iPhone 13 Pro Max
### Operating system version
iOS 16.1
### Application version
develop
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
rich text composer sometimes fullscreen mode breaks when opening it when keyboard is not showing steps to reproduce open any chat with rte enabled tap on fullscreen toggle when keyboard is not shown either dismiss it first or never bring it up outcome what did you expect full screen mode shows correctly what happened instead sometimes it shows in a broken state here is an example it s still possible to dismiss this view by dragging it to the bottom your phone model iphone pro max operating system version ios application version develop homeserver no response will you send logs no
| 1
|
17,142
| 4,141,416,590
|
IssuesEvent
|
2016-06-14 05:12:03
|
Varying-Vagrant-Vagrants/VVV
|
https://api.github.com/repos/Varying-Vagrant-Vagrants/VVV
|
opened
|
Install PHPCS globally with composer rather than git
|
documentation enhancement needs:feedback
|
Via #608, VVV doesn't necessarily need to provide PHP_Codesniffer via git clone and can instead use Composer to install it globally.
It would be interesting to hear from anyone working on the [WordPress Coding Standards](https://github.com/WordPress-Coding-Standards/WordPress-Coding-Standards) project as to whether its helpful to have the PHP_Codesniffer git repository available. /cc @westonruter, @GaryJones 😄
|
1.0
|
Install PHPCS globally with composer rather than git - Via #608, VVV doesn't necessarily need to provide PHP_Codesniffer via git clone and can instead use Composer to install it globally.
It would be interesting to hear from anyone working on the [WordPress Coding Standards](https://github.com/WordPress-Coding-Standards/WordPress-Coding-Standards) project as to whether its helpful to have the PHP_Codesniffer git repository available. /cc @westonruter, @GaryJones 😄
|
non_defect
|
install phpcs globally with composer rather than git via vvv doesn t necessarily need to provide php codesniffer via git clone and can instead use composer to install it globally it would be interesting to hear from anyone working on the project as to whether its helpful to have the php codesniffer git repository available cc westonruter garyjones 😄
| 0
|
60,029
| 14,696,612,002
|
IssuesEvent
|
2021-01-04 00:01:20
|
varkor/quiver
|
https://api.github.com/repos/varkor/quiver
|
closed
|
Doesn't make with GnuWin make for windows
|
build-system
|
http://gnuwin32.sourceforge.net/packages/make.htm
```
C:\Users\FruitfulApproach\Desktop\Quiver\quiver>make
set -e
Environment variable -e not defined
make: *** [all] Error 1
C:\Users\FruitfulApproach\Desktop\Quiver\quiver>
```
|
1.0
|
Doesn't make with GnuWin make for windows - http://gnuwin32.sourceforge.net/packages/make.htm
```
C:\Users\FruitfulApproach\Desktop\Quiver\quiver>make
set -e
Environment variable -e not defined
make: *** [all] Error 1
C:\Users\FruitfulApproach\Desktop\Quiver\quiver>
```
|
non_defect
|
doesn t make with gnuwin make for windows c users fruitfulapproach desktop quiver quiver make set e environment variable e not defined make error c users fruitfulapproach desktop quiver quiver
| 0
|
25,228
| 4,237,744,160
|
IssuesEvent
|
2016-07-05 23:08:23
|
WildBamaBoy/minecraft-comes-alive
|
https://api.github.com/repos/WildBamaBoy/minecraft-comes-alive
|
closed
|
Fences try to connect to gravestones
|
1.8.9 defect
|
Version 1.8.9-1.5.2: Fences try to connect to gravestone blocks -- they more or less succeed on the sides, but it still looks bad, and front and back are just sad.
|
1.0
|
Fences try to connect to gravestones - Version 1.8.9-1.5.2: Fences try to connect to gravestone blocks -- they more or less succeed on the sides, but it still looks bad, and front and back are just sad.
|
defect
|
fences try to connect to gravestones version fences try to connect to gravestone blocks they more or less succeed on the sides but it still looks bad and front and back are just sad
| 1
|
707,443
| 24,307,129,740
|
IssuesEvent
|
2022-09-29 18:28:21
|
authzed/spicedb
|
https://api.github.com/repos/authzed/spicedb
|
closed
|
add caveats in the WriteRelationships/ReadRelationships API
|
priority/0 urgent
|
Tracking: https://github.com/authzed/spicedb/issues/386
We should adjust the `WriteRelationships` and `ReadRelationships` API to make it possible to store caveated tuples
|
1.0
|
add caveats in the WriteRelationships/ReadRelationships API - Tracking: https://github.com/authzed/spicedb/issues/386
We should adjust the `WriteRelationships` and `ReadRelationships` API to make it possible to store caveated tuples
|
non_defect
|
add caveats in the writerelationships readrelationships api tracking we should adjust the writerelationships and readrelationships api to make it possible to store caveated tuples
| 0
|
319,451
| 23,772,939,280
|
IssuesEvent
|
2022-09-01 18:01:27
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
[Application Hosting and Deployment] Create scaffolding of Platform manual (EKS)
|
operations documentation devops infrastructure eks
|
## Description
Platform Crew developers that are deploying applications to EKS need documentation that explains how the system works along with various steps and processes involved in the application management lifecycle. The information currently exists in multiple places but could be tied together in an "easy-to-use" package.
## Background/context
- See parent epic
---
## Tasks
- [ ] Create documentation that outlines how to use the application hosting and deployment system
## Acceptance Criteria
- [ ] Documentation exists in Confluence and can be found easily
|
1.0
|
[Application Hosting and Deployment] Create scaffolding of Platform manual (EKS) - ## Description
Platform Crew developers that are deploying applications to EKS need documentation that explains how the system works along with various steps and processes involved in the application management lifecycle. The information currently exists in multiple places but could be tied together in an "easy-to-use" package.
## Background/context
- See parent epic
---
## Tasks
- [ ] Create documentation that outlines how to use the application hosting and deployment system
## Acceptance Criteria
- [ ] Documentation exists in Confluence and can be found easily
|
non_defect
|
create scaffolding of platform manual eks description platform crew developers that are deploying applications to eks need documentation that explains how the system works along with various steps and processes involved in the application management lifecycle the information currently exists in multiple places but could be tied together in an easy to use package background context see parent epic tasks create documentation that outlines how to use the application hosting and deployment system acceptance criteria documentation exists in confluence and can be found easily
| 0
|
58,117
| 16,342,511,841
|
IssuesEvent
|
2021-05-13 00:28:40
|
darshan-hpc/darshan
|
https://api.github.com/repos/darshan-hpc/darshan
|
closed
|
stdio regression tests are failing on BG/Q
|
defect
|
In GitLab by @shanedsnyder on Jul 7, 2016, 16:35
The stdio test cases used in our regression tests are working correctly on Jenkins, personal workstation, but not working correctly on the BG/Q.
Just glancing at the output, it looks like the issue may be that our wrapper functions aren't intercepting the fstream calls, but need to investigate why.
|
1.0
|
stdio regression tests are failing on BG/Q - In GitLab by @shanedsnyder on Jul 7, 2016, 16:35
The stdio test cases used in our regression tests are working correctly on Jenkins, personal workstation, but not working correctly on the BG/Q.
Just glancing at the output, it looks like the issue may be that our wrapper functions aren't intercepting the fstream calls, but need to investigate why.
|
defect
|
stdio regression tests are failing on bg q in gitlab by shanedsnyder on jul the stdio test cases used in our regression tests are working correctly on jenkins personal workstation but not working correctly on the bg q just glancing at the output it looks like the issue may be that our wrapper functions aren t intercepting the fstream calls but need to investigate why
| 1
|
44,041
| 11,935,715,310
|
IssuesEvent
|
2020-04-02 09:04:08
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
Explicitly registering Portable ClassDefinitions fails with multiple PortableFactories and overlapping classIds
|
Source: Community Team: Client Team: Core Type: Defect
|
<!--
Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently.
-->
**Describe the bug**
To work around some issues I was having with nullable Portable fields, I tried to explicitly register class definitions, following the documentation. This went well until I tried registering some classes for a second PortableFactory, that are using classIds that overlap with those of the first PortableFactory. As far as I understood from the documentation, classIds only need to be unique per PortableFactory. I then got the following exception:
```
Exception in thread "main" com.hazelcast.nio.serialization.HazelcastSerializationException: Duplicate registration found for class-id[1]!
at com.hazelcast.internal.serialization.impl.SerializationServiceV1.registerClassDefinitions(SerializationServiceV1.java:191)
at com.hazelcast.internal.serialization.impl.DefaultSerializationServiceBuilder.createSerializationService(DefaultSerializationServiceBuilder.java:300)
at com.hazelcast.internal.serialization.impl.DefaultSerializationServiceBuilder.build(DefaultSerializationServiceBuilder.java:237)
at com.hazelcast.internal.serialization.impl.DefaultSerializationServiceBuilder.build(DefaultSerializationServiceBuilder.java:55)
at com.hazelcast.instance.DefaultNodeExtension.createSerializationService(DefaultNodeExtension.java:210)
at com.hazelcast.instance.Node.<init>(Node.java:240)
at com.hazelcast.instance.HazelcastInstanceImpl.createNode(HazelcastInstanceImpl.java:161)
at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:131)
at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:229)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:208)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:158)
at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:57)
at eu.sensara.procare.residentstate.TestCase.main(TestCase.java:120)
```
**Expected behavior**
As described in the javadoc of Portable.getClassId(), I expected classIds to only be unique per PortableFactory, I expected no exception:
```
Returns class identifier for this portable class. Class ID should be unique per PortableFactory.
```
**To Reproduce**
I've reproduced the issue in this self-contained test case:
```
public class TestCase {
public static class MyPortable1 implements Portable {
public static final int ID = 1;
private String stringField;
@Override
public int getFactoryId() {
return MyPortableFactory1.ID;
}
@Override
public int getClassId() {
return ID;
}
@Override
public void writePortable(final PortableWriter writer) throws IOException {
writer.writeUTF("stringField", stringField);
}
@Override
public void readPortable(final PortableReader reader) throws IOException {
stringField = reader.readUTF("stringField");
}
public static ClassDefinition getClassDefinition() {
return new ClassDefinitionBuilder(MyPortableFactory1.ID, ID)
.addUTFField("stringField")
.build();
}
}
public static class MyPortable2 implements Portable {
public static final int ID = 1;
private int intField;
@Override
public int getFactoryId() {
return MyPortableFactory2.ID;
}
@Override
public int getClassId() {
return ID;
}
@Override
public void writePortable(final PortableWriter writer) throws IOException {
writer.writeInt("intField", intField);
}
@Override
public void readPortable(final PortableReader reader) throws IOException {
intField = reader.readInt("intField");
}
public static ClassDefinition getClassDefinition() {
return new ClassDefinitionBuilder(MyPortableFactory1.ID, ID)
.addIntField("intField")
.build();
}
}
public static class MyPortableFactory1 implements PortableFactory {
public static final int ID = 1;
@Override
public Portable create(final int classId) {
if (classId == MyPortable1.ID) {
return new MyPortable1();
}
return null;
}
}
public static class MyPortableFactory2 implements PortableFactory {
public static final int ID = 2;
@Override
public Portable create(final int classId) {
if (classId == MyPortable2.ID) {
return new MyPortable2();
}
return null;
}
}
public static void main(final String[] args) {
Config cfg = new Config();
cfg.getSerializationConfig()
.addPortableFactory(MyPortableFactory1.ID, new MyPortableFactory1())
.addPortableFactory(MyPortableFactory2.ID, new MyPortableFactory2())
.addClassDefinition(MyPortable1.getClassDefinition())
.addClassDefinition(MyPortable2.getClassDefinition());
HazelcastInstance hazelcast = Hazelcast.newHazelcastInstance(cfg);
}
}
```
**Additional context**
I've produced this issue with Hazelcast version 3.12.6, on openjdk 11.0.6.
It seems like SerializationServiceV1 creates a single HashMap that is then filled with all class definitions (of all PortableFactories), using the classId as the key. The result is that it encounters the same key for a second time and throws the HazelcastSerializationException.
I hope I am reading the documentation correctly. I was not able to find anything specifying this as a known limitation of explictly registering ClassDefinitions.
|
1.0
|
Explicitly registering Portable ClassDefinitions fails with multiple PortableFactories and overlapping classIds - <!--
Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently.
-->
**Describe the bug**
To work around some issues I was having with nullable Portable fields, I tried to explicitly register class definitions, following the documentation. This went well until I tried registering some classes for a second PortableFactory, that are using classIds that overlap with those of the first PortableFactory. As far as I understood from the documentation, classIds only need to be unique per PortableFactory. I then got the following exception:
```
Exception in thread "main" com.hazelcast.nio.serialization.HazelcastSerializationException: Duplicate registration found for class-id[1]!
at com.hazelcast.internal.serialization.impl.SerializationServiceV1.registerClassDefinitions(SerializationServiceV1.java:191)
at com.hazelcast.internal.serialization.impl.DefaultSerializationServiceBuilder.createSerializationService(DefaultSerializationServiceBuilder.java:300)
at com.hazelcast.internal.serialization.impl.DefaultSerializationServiceBuilder.build(DefaultSerializationServiceBuilder.java:237)
at com.hazelcast.internal.serialization.impl.DefaultSerializationServiceBuilder.build(DefaultSerializationServiceBuilder.java:55)
at com.hazelcast.instance.DefaultNodeExtension.createSerializationService(DefaultNodeExtension.java:210)
at com.hazelcast.instance.Node.<init>(Node.java:240)
at com.hazelcast.instance.HazelcastInstanceImpl.createNode(HazelcastInstanceImpl.java:161)
at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:131)
at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:229)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:208)
at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:158)
at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:57)
at eu.sensara.procare.residentstate.TestCase.main(TestCase.java:120)
```
**Expected behavior**
As described in the javadoc of Portable.getClassId(), I expected classIds to only be unique per PortableFactory, I expected no exception:
```
Returns class identifier for this portable class. Class ID should be unique per PortableFactory.
```
**To Reproduce**
I've reproduced the issue in this self-contained test case:
```
public class TestCase {
public static class MyPortable1 implements Portable {
public static final int ID = 1;
private String stringField;
@Override
public int getFactoryId() {
return MyPortableFactory1.ID;
}
@Override
public int getClassId() {
return ID;
}
@Override
public void writePortable(final PortableWriter writer) throws IOException {
writer.writeUTF("stringField", stringField);
}
@Override
public void readPortable(final PortableReader reader) throws IOException {
stringField = reader.readUTF("stringField");
}
public static ClassDefinition getClassDefinition() {
return new ClassDefinitionBuilder(MyPortableFactory1.ID, ID)
.addUTFField("stringField")
.build();
}
}
public static class MyPortable2 implements Portable {
public static final int ID = 1;
private int intField;
@Override
public int getFactoryId() {
return MyPortableFactory2.ID;
}
@Override
public int getClassId() {
return ID;
}
@Override
public void writePortable(final PortableWriter writer) throws IOException {
writer.writeInt("intField", intField);
}
@Override
public void readPortable(final PortableReader reader) throws IOException {
intField = reader.readInt("intField");
}
public static ClassDefinition getClassDefinition() {
return new ClassDefinitionBuilder(MyPortableFactory1.ID, ID)
.addIntField("intField")
.build();
}
}
public static class MyPortableFactory1 implements PortableFactory {
public static final int ID = 1;
@Override
public Portable create(final int classId) {
if (classId == MyPortable1.ID) {
return new MyPortable1();
}
return null;
}
}
public static class MyPortableFactory2 implements PortableFactory {
public static final int ID = 2;
@Override
public Portable create(final int classId) {
if (classId == MyPortable2.ID) {
return new MyPortable2();
}
return null;
}
}
public static void main(final String[] args) {
Config cfg = new Config();
cfg.getSerializationConfig()
.addPortableFactory(MyPortableFactory1.ID, new MyPortableFactory1())
.addPortableFactory(MyPortableFactory2.ID, new MyPortableFactory2())
.addClassDefinition(MyPortable1.getClassDefinition())
.addClassDefinition(MyPortable2.getClassDefinition());
HazelcastInstance hazelcast = Hazelcast.newHazelcastInstance(cfg);
}
}
```
**Additional context**
I've produced this issue with Hazelcast version 3.12.6, on openjdk 11.0.6.
It seems like SerializationServiceV1 creates a single HashMap that is then filled with all class definitions (of all PortableFactories), using the classId as the key. The result is that it encounters the same key for a second time and throws the HazelcastSerializationException.
I hope I am reading the documentation correctly. I was not able to find anything specifying this as a known limitation of explictly registering ClassDefinitions.
|
defect
|
explicitly registering portable classdefinitions fails with multiple portablefactories and overlapping classids thanks for reporting your issue please share with us the following information to help us resolve your issue quickly and efficiently describe the bug to work around some issues i was having with nullable portable fields i tried to explicitly register class definitions following the documentation this went well until i tried registering some classes for a second portablefactory that are using classids that overlap with those of the first portablefactory as far as i understood from the documentation classids only need to be unique per portablefactory i then got the following exception exception in thread main com hazelcast nio serialization hazelcastserializationexception duplicate registration found for class id at com hazelcast internal serialization impl registerclassdefinitions java at com hazelcast internal serialization impl defaultserializationservicebuilder createserializationservice defaultserializationservicebuilder java at com hazelcast internal serialization impl defaultserializationservicebuilder build defaultserializationservicebuilder java at com hazelcast internal serialization impl defaultserializationservicebuilder build defaultserializationservicebuilder java at com hazelcast instance defaultnodeextension createserializationservice defaultnodeextension java at com hazelcast instance node node java at com hazelcast instance hazelcastinstanceimpl createnode hazelcastinstanceimpl java at com hazelcast instance hazelcastinstanceimpl hazelcastinstanceimpl java at com hazelcast instance hazelcastinstancefactory constructhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast core hazelcast newhazelcastinstance hazelcast java at eu sensara procare residentstate testcase main testcase java expected behavior as described in the javadoc of portable getclassid i expected classids to only be unique per portablefactory i expected no exception returns class identifier for this portable class class id should be unique per portablefactory to reproduce i ve reproduced the issue in this self contained test case public class testcase public static class implements portable public static final int id private string stringfield override public int getfactoryid return id override public int getclassid return id override public void writeportable final portablewriter writer throws ioexception writer writeutf stringfield stringfield override public void readportable final portablereader reader throws ioexception stringfield reader readutf stringfield public static classdefinition getclassdefinition return new classdefinitionbuilder id id addutffield stringfield build public static class implements portable public static final int id private int intfield override public int getfactoryid return id override public int getclassid return id override public void writeportable final portablewriter writer throws ioexception writer writeint intfield intfield override public void readportable final portablereader reader throws ioexception intfield reader readint intfield public static classdefinition getclassdefinition return new classdefinitionbuilder id id addintfield intfield build public static class implements portablefactory public static final int id override public portable create final int classid if classid id return new return null public static class implements portablefactory public static final int id override public portable create final int classid if classid id return new return null public static void main final string args config cfg new config cfg getserializationconfig addportablefactory id new addportablefactory id new addclassdefinition getclassdefinition addclassdefinition getclassdefinition hazelcastinstance hazelcast hazelcast newhazelcastinstance cfg additional context i ve produced this issue with hazelcast version on openjdk it seems like creates a single hashmap that is then filled with all class definitions of all portablefactories using the classid as the key the result is that it encounters the same key for a second time and throws the hazelcastserializationexception i hope i am reading the documentation correctly i was not able to find anything specifying this as a known limitation of explictly registering classdefinitions
| 1
|
71,212
| 23,491,116,552
|
IssuesEvent
|
2022-08-17 18:50:45
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
508-defect-3 [COGNITION]: Consider updating 24/7 HTML for screen reader understanding (VAMC)
|
frontend 508/Accessibility vsa 508-defect-3 Facilities 508-issue-cognition frontend-vamc vamc-pages
|
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Some pages are using a shorthand pattern like **Mon:** 24/7 to represent Monday, 24 hours a day, 7 days a week. This is an implicit relationship, but is not made clear to assistive technology. I'd like to consider updating this markup to better serve screen reader users. Code snippet below.
The announcement of the full day name has been moved into #21731 as it can be handled from the template. This issue is now limited to handling 24/7 handling.
## Type of Issue
template
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
* [ ] Screen reader announces the hours of the day and days of the week
## Environment
* MacOS Mojave
* Safari
* VoiceOver
## Possible Fixes (optional)
```html
<ul class="vads-u-flex--1 va-c-facility-hours-list vads-u-margin-top--0 vads-u-margin-bottom--1 small-screen:vads-u-margin-bottom--0 vads-u-margin-right--3">
<li>
<b class="abbrv-day">
<span aria-hidden="true">Mon:</span>
<span class="vads-u-visibility--screen-reader">Monday</span>
</b>
<span aria-hidden="true">24/7</span>
<span class="vads-u-visibility--screen-reader">Twenty-four hours a day. Seven days a week.</span>
</li>
</ul>
```
## WCAG or Vendor Guidance (optional)
* [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->
<img width="657" alt="Screen Shot 2020-02-19 at 4 31 54 PM" src="https://user-images.githubusercontent.com/934879/74883339-5e151200-5336-11ea-8723-743807338d1e.png">
|
1.0
|
508-defect-3 [COGNITION]: Consider updating 24/7 HTML for screen reader understanding (VAMC) - ## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Some pages are using a shorthand pattern like **Mon:** 24/7 to represent Monday, 24 hours a day, 7 days a week. This is an implicit relationship, but is not made clear to assistive technology. I'd like to consider updating this markup to better serve screen reader users. Code snippet below.
The announcement of the full day name has been moved into #21731 as it can be handled from the template. This issue is now limited to handling 24/7 handling.
## Type of Issue
template
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
* [ ] Screen reader announces the hours of the day and days of the week
## Environment
* MacOS Mojave
* Safari
* VoiceOver
## Possible Fixes (optional)
```html
<ul class="vads-u-flex--1 va-c-facility-hours-list vads-u-margin-top--0 vads-u-margin-bottom--1 small-screen:vads-u-margin-bottom--0 vads-u-margin-right--3">
<li>
<b class="abbrv-day">
<span aria-hidden="true">Mon:</span>
<span class="vads-u-visibility--screen-reader">Monday</span>
</b>
<span aria-hidden="true">24/7</span>
<span class="vads-u-visibility--screen-reader">Twenty-four hours a day. Seven days a week.</span>
</li>
</ul>
```
## WCAG or Vendor Guidance (optional)
* [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->
<img width="657" alt="Screen Shot 2020-02-19 at 4 31 54 PM" src="https://user-images.githubusercontent.com/934879/74883339-5e151200-5336-11ea-8723-743807338d1e.png">
|
defect
|
defect consider updating html for screen reader understanding vamc description some pages are using a shorthand pattern like mon to represent monday hours a day days a week this is an implicit relationship but is not made clear to assistive technology i d like to consider updating this markup to better serve screen reader users code snippet below the announcement of the full day name has been moved into as it can be handled from the template this issue is now limited to handling handling type of issue template point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact trevor acceptance criteria screen reader announces the hours of the day and days of the week environment macos mojave safari voiceover possible fixes optional html mon monday twenty four hours a day seven days a week wcag or vendor guidance optional screenshots or trace logs img width alt screen shot at pm src
| 1
|
11,429
| 2,651,459,151
|
IssuesEvent
|
2015-03-16 11:40:31
|
douglasdrumond/cloaked-computing-machine
|
https://api.github.com/repos/douglasdrumond/cloaked-computing-machine
|
opened
|
[CLOSED] resizing doesn't always update main pane
|
auto-migrated Priority-Medium Type-Defect
|
<a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Mar 16, 2015 at 09:15 GMT_
_Originally opened as https://github.com/douglasdrumond/macvim/issues/15_
----
```
resizing doesn't always keep up. The main content doesn't always notice window
resizes.
Very nice implementation though - my favorite yet.
```
Original issue reported on code.google.com by `georgeha...@gmail.com` on 15 Sep 2007 at 11:06
|
1.0
|
[CLOSED] resizing doesn't always update main pane - <a href="https://github.com/GoogleCodeExporter"><img src="https://avatars.githubusercontent.com/u/9614759?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [GoogleCodeExporter](https://github.com/GoogleCodeExporter)**
_Monday Mar 16, 2015 at 09:15 GMT_
_Originally opened as https://github.com/douglasdrumond/macvim/issues/15_
----
```
resizing doesn't always keep up. The main content doesn't always notice window
resizes.
Very nice implementation though - my favorite yet.
```
Original issue reported on code.google.com by `georgeha...@gmail.com` on 15 Sep 2007 at 11:06
|
defect
|
resizing doesn t always update main pane issue by monday mar at gmt originally opened as resizing doesn t always keep up the main content doesn t always notice window resizes very nice implementation though my favorite yet original issue reported on code google com by georgeha gmail com on sep at
| 1
|
48,906
| 7,463,589,916
|
IssuesEvent
|
2018-04-01 07:46:34
|
tugcanolgun/SWE573
|
https://api.github.com/repos/tugcanolgun/SWE573
|
closed
|
Prepare the project plan
|
Documentation Low Priority
|
Prepare the project plan on a tool to be able to see the process while doing the project.
|
1.0
|
Prepare the project plan - Prepare the project plan on a tool to be able to see the process while doing the project.
|
non_defect
|
prepare the project plan prepare the project plan on a tool to be able to see the process while doing the project
| 0
|
57,376
| 14,102,503,949
|
IssuesEvent
|
2020-11-06 08:51:50
|
smartstore/SmartStoreNET
|
https://api.github.com/repos/smartstore/SmartStoreNET
|
closed
|
Page Builder: Add Content slider as block
|
pagebuilder
|
It would be great if you can add a Content Slider from the[ CMS Page Builder](http://community.smartstore.com/index.php?/files/file/163-cms-page-builder/) plugin as a block to your Page Builder story. To extend this feature it would be awesome if you can also add a Pagebuilder Story Slider as a block. Meaning that you can create stories that you can then bundle in a slider.
|
1.0
|
Page Builder: Add Content slider as block - It would be great if you can add a Content Slider from the[ CMS Page Builder](http://community.smartstore.com/index.php?/files/file/163-cms-page-builder/) plugin as a block to your Page Builder story. To extend this feature it would be awesome if you can also add a Pagebuilder Story Slider as a block. Meaning that you can create stories that you can then bundle in a slider.
|
non_defect
|
page builder add content slider as block it would be great if you can add a content slider from the plugin as a block to your page builder story to extend this feature it would be awesome if you can also add a pagebuilder story slider as a block meaning that you can create stories that you can then bundle in a slider
| 0
|
138,136
| 11,184,727,018
|
IssuesEvent
|
2019-12-31 19:51:12
|
eventespresso/eea-multisite
|
https://api.github.com/repos/eventespresso/eea-multisite
|
reopened
|
Fix Unit test failures
|
category:unit-tests status:stale type:bug 🐞
|
<!--
BEFORE POSTING YOUR ISSUE:
- These comments won't show up when you submit the issue.
- Please ensure that what you are reporting is specific to this project.
- Try to add as much detail as possible. Be specific!
- Make sure you read the README.md for the project regarding posting issues.
- Search this repository for issues and pull requests and whether it has been fixed or reported already.
- Ensure you are using the latest code before reporting bugs (unless you are reporting an issue disovered in a branch).
- Disable all plugins and switch to a default theme to ensure its not a plugin/theme conflict issue.
- To report a security issue, please visit this page: https://eventespresso.com/report-a-security-vulnerability/
-->
## Issue Overview
<!-- Describe what this issue is about. -->
There are a bunch of unit test failures. I only noticed them today (after the release of EE core 4.10) but they seem to have existed since before then. We just hadn't noticed them from the nightly builds. Eg https://travis-ci.com/eventespresso/eea-multisite/jobs/218461559
Fixing these would be nice in order to confirm there's no actual problem, and to make it more clear if there is a problem introduced in the future.
## Bug report or feature request?
* [x] Bug
* [ ] Feature
* [ ] Neither
## Steps to Reproduce (for bugs)
<!-- If possible provide any links to a live example, or an unambiguous set of steps to reproduce this bug -->
<!-- Feel free to include code to reproduce if relevant. -->
Run the multisite tests locally (or refer to the latest nightly builds on Circle). There are a lot of unit test failures.
|
1.0
|
Fix Unit test failures - <!--
BEFORE POSTING YOUR ISSUE:
- These comments won't show up when you submit the issue.
- Please ensure that what you are reporting is specific to this project.
- Try to add as much detail as possible. Be specific!
- Make sure you read the README.md for the project regarding posting issues.
- Search this repository for issues and pull requests and whether it has been fixed or reported already.
- Ensure you are using the latest code before reporting bugs (unless you are reporting an issue disovered in a branch).
- Disable all plugins and switch to a default theme to ensure its not a plugin/theme conflict issue.
- To report a security issue, please visit this page: https://eventespresso.com/report-a-security-vulnerability/
-->
## Issue Overview
<!-- Describe what this issue is about. -->
There are a bunch of unit test failures. I only noticed them today (after the release of EE core 4.10) but they seem to have existed since before then. We just hadn't noticed them from the nightly builds. Eg https://travis-ci.com/eventespresso/eea-multisite/jobs/218461559
Fixing these would be nice in order to confirm there's no actual problem, and to make it more clear if there is a problem introduced in the future.
## Bug report or feature request?
* [x] Bug
* [ ] Feature
* [ ] Neither
## Steps to Reproduce (for bugs)
<!-- If possible provide any links to a live example, or an unambiguous set of steps to reproduce this bug -->
<!-- Feel free to include code to reproduce if relevant. -->
Run the multisite tests locally (or refer to the latest nightly builds on Circle). There are a lot of unit test failures.
|
non_defect
|
fix unit test failures before posting your issue these comments won t show up when you submit the issue please ensure that what you are reporting is specific to this project try to add as much detail as possible be specific make sure you read the readme md for the project regarding posting issues search this repository for issues and pull requests and whether it has been fixed or reported already ensure you are using the latest code before reporting bugs unless you are reporting an issue disovered in a branch disable all plugins and switch to a default theme to ensure its not a plugin theme conflict issue to report a security issue please visit this page issue overview there are a bunch of unit test failures i only noticed them today after the release of ee core but they seem to have existed since before then we just hadn t noticed them from the nightly builds eg fixing these would be nice in order to confirm there s no actual problem and to make it more clear if there is a problem introduced in the future bug report or feature request bug feature neither steps to reproduce for bugs run the multisite tests locally or refer to the latest nightly builds on circle there are a lot of unit test failures
| 0
|
8,484
| 3,184,556,496
|
IssuesEvent
|
2015-09-27 13:55:31
|
owncloud/gallery
|
https://api.github.com/repos/owncloud/gallery
|
closed
|
v13.0.0 - pictures on remote shares are not shown
|
documentation enhancement
|
## Expected behavior
The galleryplus app should also show pictures on remote shares, like WebDAV shares etc.
## Current behavior
Only locally stored pictures are shown in the galleryplus app. Several log messages are written but nothing seems to point to the root cause of the problem. I wonder if the app supports showing pictures on remote shares, like WebDAV shares, at all?
```
...,"app":"core","message":"DB prepare : SELECT `appid`, `configvalue` FROM `oc_appconfig` WHERE `configkey` = 'installed_version'","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `gid` FROM `oc_groups` WHERE `gid` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'admin', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `uid`, `displayname` FROM `oc_users` WHERE LOWER(`uid`) = LOWER(?)","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
....,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
....,"app":"core","message":"DB prepare : SELECT `uid` FROM `oc_group_user` WHERE `gid` = ? AND `uid` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'admin', 1 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id`, `displayname`, `description`, `ctag` AS `lastmodified`, `userid` AS `owner`, `uri` FROM `oc_contacts_addressbooks` WHERE `userid` = ? ORDER BY `displayname`","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `gid` FROM `oc_group_user` WHERE `uid` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT * FROM `oc_share` WHERE `item_type` = ? AND ((`share_type` in (?, ?) AND `share_with` = ?) OR (`share_type` = ? AND `share_with` IN (?,?,?)) ) AND `uid_owner` != ? ORDER BY `oc_share`.`id` ASC","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'addressbook', 1 => 0, 2 => 2, 3 => 'juergen', 4 => 1, 5 => 'Eltern', 6 => 'Familie', 7 => 'admin', 8 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `numeric_id` FROM `oc_storages` WHERE `id` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'local::\/var\/owncloud\/data\/juergen\/', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `gid` FROM `oc_group_user` WHERE `uid` =?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `oc_share`.`id`, `item_type`, `item_source`, `item_target`,`oc_share`.`parent`, `share_type`, `share_with`, `uid_owner`,`file_source`, `path`, `file_target`, `oc_share`.`permissions`,`stime`, `expiration`, `token`, `storage`, `mail_send`,`oc_storages`.`id` AS `storage_id` FROM `oc_share` INNER JOIN `oc_filecache` ON `file_source` = `oc_filecache`.`fileid` AND `file_target` IS NOT NULL INNER JOIN `oc_storages` ON `numeric_id` = `oc_filecache`.`storage` AND ((`share_type` in (?, ?) AND `share_with` = ?) OR (`share_type` = ? AND `share_with` IN (?,?,?)) ) AND `uid_owner` != ? ORDER BY `oc_share`.`id` ASC","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
```
## Steps to reproduce
1. Open the galleryplus app
2. check if pictures on remote shares are shown.
## Environment
#### Server Configuration
OS: Linux 3.2.67
Web server: Apache2 2.4.12
Database: MySQL 5.1.73
PHP version: 5.4.42
ownCloud version: 8.0.5
GalleryPlus app version: 13.0.0
#### Client Configuration
Operating system: Windows 7
Firefox version: 39.0
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/24508941-v13-0-0-pictures-on-remote-shares-are-not-shown?utm_campaign=plugin&utm_content=tracker%2F9328526&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F9328526&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
v13.0.0 - pictures on remote shares are not shown - ## Expected behavior
The galleryplus app should also show pictures on remote shares, like WebDAV shares etc.
## Current behavior
Only locally stored pictures are shown in the galleryplus app. Several log messages are written but nothing seems to point to the root cause of the problem. I wonder if the app supports showing pictures on remote shares, like WebDAV shares, at all?
```
...,"app":"core","message":"DB prepare : SELECT `appid`, `configvalue` FROM `oc_appconfig` WHERE `configkey` = 'installed_version'","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `gid` FROM `oc_groups` WHERE `gid` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'admin', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `uid`, `displayname` FROM `oc_users` WHERE LOWER(`uid`) = LOWER(?)","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
....,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
....,"app":"core","message":"DB prepare : SELECT `uid` FROM `oc_group_user` WHERE `gid` = ? AND `uid` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'admin', 1 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id`, `displayname`, `description`, `ctag` AS `lastmodified`, `userid` AS `owner`, `uri` FROM `oc_contacts_addressbooks` WHERE `userid` = ? ORDER BY `displayname`","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `gid` FROM `oc_group_user` WHERE `uid` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT * FROM `oc_share` WHERE `item_type` = ? AND ((`share_type` in (?, ?) AND `share_with` = ?) OR (`share_type` = ? AND `share_with` IN (?,?,?)) ) AND `uid_owner` != ? ORDER BY `oc_share`.`id` ASC","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'addressbook', 1 => 0, 2 => 2, 3 => 'juergen', 4 => 1, 5 => 'Eltern', 6 => 'Familie', 7 => 'admin', 8 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `id` FROM `oc_jobs` WHERE `class` = ? AND `argument` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `numeric_id` FROM `oc_storages` WHERE `id` = ?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'local::\/var\/owncloud\/data\/juergen\/', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `gid` FROM `oc_group_user` WHERE `uid` =?","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB execute with arguments : array ( 0 => 'juergen', )","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
...,"app":"core","message":"DB prepare : SELECT `oc_share`.`id`, `item_type`, `item_source`, `item_target`,`oc_share`.`parent`, `share_type`, `share_with`, `uid_owner`,`file_source`, `path`, `file_target`, `oc_share`.`permissions`,`stime`, `expiration`, `token`, `storage`, `mail_send`,`oc_storages`.`id` AS `storage_id` FROM `oc_share` INNER JOIN `oc_filecache` ON `file_source` = `oc_filecache`.`fileid` AND `file_target` IS NOT NULL INNER JOIN `oc_storages` ON `numeric_id` = `oc_filecache`.`storage` AND ((`share_type` in (?, ?) AND `share_with` = ?) OR (`share_type` = ? AND `share_with` IN (?,?,?)) ) AND `uid_owner` != ? ORDER BY `oc_share`.`id` ASC","level":0,"time":"2015-07-17 18:31:34+02:00","method":"GET","url":"\/owncloud\/index.php\/apps\/galleryplus\/"}
```
## Steps to reproduce
1. Open the galleryplus app
2. check if pictures on remote shares are shown.
## Environment
#### Server Configuration
OS: Linux 3.2.67
Web server: Apache2 2.4.12
Database: MySQL 5.1.73
PHP version: 5.4.42
ownCloud version: 8.0.5
GalleryPlus app version: 13.0.0
#### Client Configuration
Operating system: Windows 7
Firefox version: 39.0
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/24508941-v13-0-0-pictures-on-remote-shares-are-not-shown?utm_campaign=plugin&utm_content=tracker%2F9328526&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F9328526&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_defect
|
pictures on remote shares are not shown expected behavior the galleryplus app should also show pictures on remote shares like webdav shares etc current behavior only locally stored pictures are shown in the galleryplus app several log messages are written but nothing seems to point to the root cause of the problem i wonder if the app supports showing pictures on remote shares like webdav shares at all app core message db prepare select appid configvalue from oc appconfig where configkey installed version level time method get url owncloud index php apps galleryplus app core message db execute with arguments array level time method get url owncloud index php apps galleryplus app core message db prepare select gid from oc groups where gid level time method get url owncloud index php apps galleryplus app core message db execute with arguments array admin level time method get url owncloud index php apps galleryplus app core message db prepare select uid displayname from oc users where lower uid lower level time method get url owncloud index php apps galleryplus app core message db execute with arguments array juergen level time method get url owncloud index php apps galleryplus app core message db prepare select uid from oc group user where gid and uid level time method get url owncloud index php apps galleryplus app core message db execute with arguments array admin juergen level time method get url owncloud index php apps galleryplus app core message db prepare select id from oc jobs where class and argument level time method get url owncloud index php apps galleryplus app core message db prepare select id from oc jobs where class and argument level time method get url owncloud index php apps galleryplus app core message db prepare select id from oc jobs where class and argument level time method get url owncloud index php apps galleryplus app core message db prepare select id displayname description ctag as lastmodified userid as owner uri from oc contacts addressbooks where userid order by displayname level time method get url owncloud index php apps galleryplus app core message db execute with arguments array juergen level time method get url owncloud index php apps galleryplus app core message db prepare select gid from oc group user where uid level time method get url owncloud index php apps galleryplus app core message db execute with arguments array juergen level time method get url owncloud index php apps galleryplus app core message db prepare select from oc share where item type and share type in and share with or share type and share with in and uid owner order by oc share id asc level time method get url owncloud index php apps galleryplus app core message db execute with arguments array addressbook juergen eltern familie admin juergen level time method get url owncloud index php apps galleryplus app core message db prepare select id from oc jobs where class and argument level time method get url owncloud index php apps galleryplus app core message db prepare select numeric id from oc storages where id level time method get url owncloud index php apps galleryplus app core message db execute with arguments array local var owncloud data juergen level time method get url owncloud index php apps galleryplus app core message db prepare select gid from oc group user where uid level time method get url owncloud index php apps galleryplus app core message db execute with arguments array juergen level time method get url owncloud index php apps galleryplus app core message db prepare select oc share id item type item source item target oc share parent share type share with uid owner file source path file target oc share permissions stime expiration token storage mail send oc storages id as storage id from oc share inner join oc filecache on file source oc filecache fileid and file target is not null inner join oc storages on numeric id oc filecache storage and share type in and share with or share type and share with in and uid owner order by oc share id asc level time method get url owncloud index php apps galleryplus steps to reproduce open the galleryplus app check if pictures on remote shares are shown environment server configuration os linux web server database mysql php version owncloud version galleryplus app version client configuration operating system windows firefox version want to back this issue we accept bounties via
| 0
|
11,371
| 2,649,801,846
|
IssuesEvent
|
2015-03-15 09:26:22
|
stumitchell/pulp-or
|
https://api.github.com/repos/stumitchell/pulp-or
|
closed
|
Some iterables don't work in LpAffineExpression's constructor or addInplace methods
|
auto-migrated Priority-Medium Type-Defect
|
```
At two places in LpAffineExpression's source, the expression
isinstance(e, list) or isinstance(e, GeneratorType)
is used to test for an iterable. In Python 3, however, there are numerous
iterables which are neither lists nor generators (for example, the keys() or
values() views of dictionaries). Consequently, code like
prob += lpSum(dct.values())
fails. The attached patch fixes the issue by using isinstance(e, Iterable)
instead (with collections.Iterable). I am not sure if this runs in python2.5,
but at least python2.6 should be fine.
```
Original issue reported on code.google.com by `superm...@googlemail.com` on 12 Jan 2015 at 1:17
Attachments:
* [fixIterables.patch](https://storage.googleapis.com/google-code-attachments/pulp-or/issue-69/comment-0/fixIterables.patch)
|
1.0
|
Some iterables don't work in LpAffineExpression's constructor or addInplace methods - ```
At two places in LpAffineExpression's source, the expression
isinstance(e, list) or isinstance(e, GeneratorType)
is used to test for an iterable. In Python 3, however, there are numerous
iterables which are neither lists nor generators (for example, the keys() or
values() views of dictionaries). Consequently, code like
prob += lpSum(dct.values())
fails. The attached patch fixes the issue by using isinstance(e, Iterable)
instead (with collections.Iterable). I am not sure if this runs in python2.5,
but at least python2.6 should be fine.
```
Original issue reported on code.google.com by `superm...@googlemail.com` on 12 Jan 2015 at 1:17
Attachments:
* [fixIterables.patch](https://storage.googleapis.com/google-code-attachments/pulp-or/issue-69/comment-0/fixIterables.patch)
|
defect
|
some iterables don t work in lpaffineexpression s constructor or addinplace methods at two places in lpaffineexpression s source the expression isinstance e list or isinstance e generatortype is used to test for an iterable in python however there are numerous iterables which are neither lists nor generators for example the keys or values views of dictionaries consequently code like prob lpsum dct values fails the attached patch fixes the issue by using isinstance e iterable instead with collections iterable i am not sure if this runs in but at least should be fine original issue reported on code google com by superm googlemail com on jan at attachments
| 1
|
71,439
| 23,628,628,577
|
IssuesEvent
|
2022-08-25 07:22:33
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
p-radioButton: Set preselection with a key instead of an object
|
defect
|
### Describe the bug
Need to set a default value for `p-radioButton` with a specific key of provided values. but now we can preselect the component just by passing whole object. The ideal is to behave like p-dropdown.
### Environment
windows
### Reproducer
_No response_
### Angular version
14.1.0
### PrimeNG version
14.0.0
### Build / Runtime
Angular CLI App
### Language
TypeScript
### Node version (for AoT issues node --version)
16.6.4
### Browser(s)
_No response_
### Steps to reproduce the behavior
_No response_
### Expected behavior
_No response_
|
1.0
|
p-radioButton: Set preselection with a key instead of an object - ### Describe the bug
Need to set a default value for `p-radioButton` with a specific key of provided values. but now we can preselect the component just by passing whole object. The ideal is to behave like p-dropdown.
### Environment
windows
### Reproducer
_No response_
### Angular version
14.1.0
### PrimeNG version
14.0.0
### Build / Runtime
Angular CLI App
### Language
TypeScript
### Node version (for AoT issues node --version)
16.6.4
### Browser(s)
_No response_
### Steps to reproduce the behavior
_No response_
### Expected behavior
_No response_
|
defect
|
p radiobutton set preselection with a key instead of an object describe the bug need to set a default value for p radiobutton with a specific key of provided values but now we can preselect the component just by passing whole object the ideal is to behave like p dropdown environment windows reproducer no response angular version primeng version build runtime angular cli app language typescript node version for aot issues node version browser s no response steps to reproduce the behavior no response expected behavior no response
| 1
|
99,605
| 4,057,294,082
|
IssuesEvent
|
2016-05-24 21:34:26
|
smartchicago/chicago-early-learning
|
https://api.github.com/repos/smartchicago/chicago-early-learning
|
closed
|
Different zoom level for zip code and community area searches
|
High Priority
|
We need to look at having different zoom level for address searches vs zip code or community area searches. Parents and Early Learning providers are not finding the appropriate location because the pinpoint is at the center of the zip code/community area and does not take the shape into account.
This came up specifically with Stock Elementary. If you type in "Edison Park" it does not appear because of the current zoom level:
<img width="1229" alt="screen shot 2016-05-16 at 2 54 23 pm" src="https://cloud.githubusercontent.com/assets/5550969/15301988/1bf7f01e-1b76-11e6-80f7-bf0448ab4eb4.png">
Stock Elementary is a little further north:
<img width="1226" alt="screen shot 2016-05-16 at 2 54 55 pm" src="https://cloud.githubusercontent.com/assets/5550969/15302002/31bf7250-1b76-11e6-81ce-336091831894.png">
Note: The current zoom level is very important for address searches because we heard from our partners that parents weren't getting locations close enough to the address they searched for. This relates to the work in this milestone: https://github.com/smartchicago/chicago-early-learning/issues?q=milestone%3A%22Improve+Search+Function%22
|
1.0
|
Different zoom level for zip code and community area searches - We need to look at having different zoom level for address searches vs zip code or community area searches. Parents and Early Learning providers are not finding the appropriate location because the pinpoint is at the center of the zip code/community area and does not take the shape into account.
This came up specifically with Stock Elementary. If you type in "Edison Park" it does not appear because of the current zoom level:
<img width="1229" alt="screen shot 2016-05-16 at 2 54 23 pm" src="https://cloud.githubusercontent.com/assets/5550969/15301988/1bf7f01e-1b76-11e6-80f7-bf0448ab4eb4.png">
Stock Elementary is a little further north:
<img width="1226" alt="screen shot 2016-05-16 at 2 54 55 pm" src="https://cloud.githubusercontent.com/assets/5550969/15302002/31bf7250-1b76-11e6-81ce-336091831894.png">
Note: The current zoom level is very important for address searches because we heard from our partners that parents weren't getting locations close enough to the address they searched for. This relates to the work in this milestone: https://github.com/smartchicago/chicago-early-learning/issues?q=milestone%3A%22Improve+Search+Function%22
|
non_defect
|
different zoom level for zip code and community area searches we need to look at having different zoom level for address searches vs zip code or community area searches parents and early learning providers are not finding the appropriate location because the pinpoint is at the center of the zip code community area and does not take the shape into account this came up specifically with stock elementary if you type in edison park it does not appear because of the current zoom level img width alt screen shot at pm src stock elementary is a little further north img width alt screen shot at pm src note the current zoom level is very important for address searches because we heard from our partners that parents weren t getting locations close enough to the address they searched for this relates to the work in this milestone
| 0
|
5,708
| 5,118,338,034
|
IssuesEvent
|
2017-01-08 03:55:38
|
angular/universal
|
https://api.github.com/repos/angular/universal
|
closed
|
perf(benchmarks): create benchmark comparing rendering on server and client
|
comp: universal effort2: moderate (half-day) P1: required type: performance
|
The goal is to determine differences in user-perceived render time, which can be referenced against common device and network capabilities to determine how the two approaches (pre-or-post-render) compare under different conditions.
There should be benchmarks of applications with different UI and data characteristics, comparing rendering just in the UI, and pre-rendering from the server. Over time, as more optimizations are available (such as inlining data and assets, service worker integration, etc), benchmarks should compare against optimized and unoptimized approaches as well.
Benchpress may be the best tool, but may need some features to support this.
The first benchmark could be based on the [angular/universal-starter](https://github.com/angular/universal-starter) app.
- [ ] Total req/res time to pre-render page and load all assets
- [ ] Total time to load and render page in client without pre-rendering, including load time
Ideally, test with these variables, using mobile emulation features in chrome devtools:
- [ ] Simulated connection quality
- [ ] Simulated hardware
TODO:
- Determine how to analyze and report on data, to segment by device processing power and network latency/throughput.
- Determine whether benchpress is the best tool, and if so, what modifications need to be made.
|
True
|
perf(benchmarks): create benchmark comparing rendering on server and client - The goal is to determine differences in user-perceived render time, which can be referenced against common device and network capabilities to determine how the two approaches (pre-or-post-render) compare under different conditions.
There should be benchmarks of applications with different UI and data characteristics, comparing rendering just in the UI, and pre-rendering from the server. Over time, as more optimizations are available (such as inlining data and assets, service worker integration, etc), benchmarks should compare against optimized and unoptimized approaches as well.
Benchpress may be the best tool, but may need some features to support this.
The first benchmark could be based on the [angular/universal-starter](https://github.com/angular/universal-starter) app.
- [ ] Total req/res time to pre-render page and load all assets
- [ ] Total time to load and render page in client without pre-rendering, including load time
Ideally, test with these variables, using mobile emulation features in chrome devtools:
- [ ] Simulated connection quality
- [ ] Simulated hardware
TODO:
- Determine how to analyze and report on data, to segment by device processing power and network latency/throughput.
- Determine whether benchpress is the best tool, and if so, what modifications need to be made.
|
non_defect
|
perf benchmarks create benchmark comparing rendering on server and client the goal is to determine differences in user perceived render time which can be referenced against common device and network capabilities to determine how the two approaches pre or post render compare under different conditions there should be benchmarks of applications with different ui and data characteristics comparing rendering just in the ui and pre rendering from the server over time as more optimizations are available such as inlining data and assets service worker integration etc benchmarks should compare against optimized and unoptimized approaches as well benchpress may be the best tool but may need some features to support this the first benchmark could be based on the app total req res time to pre render page and load all assets total time to load and render page in client without pre rendering including load time ideally test with these variables using mobile emulation features in chrome devtools simulated connection quality simulated hardware todo determine how to analyze and report on data to segment by device processing power and network latency throughput determine whether benchpress is the best tool and if so what modifications need to be made
| 0
|
457,197
| 13,153,072,229
|
IssuesEvent
|
2020-08-10 01:46:31
|
kubesphere/kubesphere
|
https://api.github.com/repos/kubesphere/kubesphere
|
closed
|
Workspace/Failed to create a workspace role with the same name as a workspace role in another workspace
|
area/console kind/bug priority/medium
|
Describe the Bug
Failed to create a workspace role with the same name as a workspace role in another workspace
Versions Used
KubeSphere:3.0.0
Environment
testing env
http://139.198.12.26:30884/
How To Reproduce
Steps to reproduce the behavior:
1.There is an workspace A, it has a workspace role named 'test-wx'
2.create a workspace role named 'test-wx' in workspace 'B'
Expected behavior
created successfully
Actual behavior
failed

/kind bug
/area console
/assign @wansir
/milestone 3.0.0
/priority medium
|
1.0
|
Workspace/Failed to create a workspace role with the same name as a workspace role in another workspace - Describe the Bug
Failed to create a workspace role with the same name as a workspace role in another workspace
Versions Used
KubeSphere:3.0.0
Environment
testing env
http://139.198.12.26:30884/
How To Reproduce
Steps to reproduce the behavior:
1.There is an workspace A, it has a workspace role named 'test-wx'
2.create a workspace role named 'test-wx' in workspace 'B'
Expected behavior
created successfully
Actual behavior
failed

/kind bug
/area console
/assign @wansir
/milestone 3.0.0
/priority medium
|
non_defect
|
workspace failed to create a workspace role with the same name as a workspace role in another workspace describe the bug failed to create a workspace role with the same name as a workspace role in another workspace versions used kubesphere environment testing env how to reproduce steps to reproduce the behavior there is an workspace a it has a workspace role named test wx create a workspace role named test wx in workspace b expected behavior created successfully actual behavior failed kind bug area console assign wansir milestone priority medium
| 0
|
337,891
| 24,561,187,939
|
IssuesEvent
|
2022-10-12 20:27:23
|
saltstack/salt
|
https://api.github.com/repos/saltstack/salt
|
closed
|
[DOCS] HA Doc needs update
|
Documentation severity-medium help-wanted good first issue time-estimate-quick-fix
|
**Description**
https://docs.saltproject.io/en/latest/topics/highavailability/index.html
In the above article we mention the options <pre>master_shuffle</pre>. According to the documentation for the minion configuration here, https://docs.saltproject.io/en/latest/ref/configuration/minion.html#master-shuffle, this option has been deprecated in favor of [random_master](https://docs.saltproject.io/en/latest/ref/configuration/minion.html#random-master).
**Suggested Fix**
Update the doc to point to random_master with relevant changes in functionality
**Type of documentation**
Salt documentation
**Location or format of documentation**
https://docs.saltproject.io/en/latest/topics/highavailability/index.html
|
1.0
|
[DOCS] HA Doc needs update - **Description**
https://docs.saltproject.io/en/latest/topics/highavailability/index.html
In the above article we mention the options <pre>master_shuffle</pre>. According to the documentation for the minion configuration here, https://docs.saltproject.io/en/latest/ref/configuration/minion.html#master-shuffle, this option has been deprecated in favor of [random_master](https://docs.saltproject.io/en/latest/ref/configuration/minion.html#random-master).
**Suggested Fix**
Update the doc to point to random_master with relevant changes in functionality
**Type of documentation**
Salt documentation
**Location or format of documentation**
https://docs.saltproject.io/en/latest/topics/highavailability/index.html
|
non_defect
|
ha doc needs update description in the above article we mention the options master shuffle according to the documentation for the minion configuration here this option has been deprecated in favor of suggested fix update the doc to point to random master with relevant changes in functionality type of documentation salt documentation location or format of documentation
| 0
|
77,243
| 26,875,296,280
|
IssuesEvent
|
2023-02-05 00:13:55
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
BUG: conda 23.1.0 environment creation fails on windows 11
|
defect
|
### Describe your issue.
I am running into issues with conda setup as described in the [environment.yml file](https://github.com/scipy/scipy/blob/main/environment.yml). This is my first time trying to setup the scipy build environment, so I am not sure yet if this is something specific to me or a larger issue. My main clue at the moment is that the error message I am getting seems to imply several conflicts in the specified yml package version numbers - which I don't think I have control over locally.
Versions for reference:
scipy source commit ddc0d31b1
anaconda 23.1.0
Windows 11
The output under the scipy/numpy/Python version information block below was generated from the base Anaconda interpreter on my machine.
### Reproducing Code Example
```python
conda env create -f .\environment.yml
```
### Error message
```shell
Collecting package metadata (repodata.json): done
Solving environment: \
Found conflicts! Looking for incompatible packages.
This can take several minutes. Press CTRL-C to abort.
Examining pytest: 16%|███████████▋ | 6/38 [00:29<01:10, 2.22s/it]-Examining pydata-sphinx-theme==0.9.0: 45%|███████████████████████▋ | 17/38 [02:18<03:12, 9.18s/it]\Examining click: 63%|██████████████████████████████████████████████▋ | 24/38 [02:24<00:39, 2.81s/it]-Examining setuptools[version='<60.0']: 79%|█████████████████████████████████████████ | 30/38 [03:00<01:11, 8.94s/it]\Examining conflict for python setuptools ruff numpy click pybind11 ipython cython pytest asv pythran matplotlib: 45%|▍| 17/38 [|Examining conflict for sphinx pooch pydata-sphinx-theme pytest matplotlib: 61%|█████████▋ | 23/38 [02:31<00:54, 3.65s/it]/Examining conflict for pytest-cov sphinx-design sphinx python numpydoc pytest-xdist click ipython rich-click pytest pydata-sphin-
failed |
Solving environment: \
Found conflicts! Looking for incompatible packages.
This can take several minutes. Press CTRL-C to abort.
Examining pytest: 18%|█████████████▋ | 7/38 [00:37<03:55, 7.59s/it]|
Examining matplotlib: 24%|████████████████▌ | 9/38 [01:28<06:17, 13.03s/it]-Examining ipython: 37%|██████████████████████████▌ | 14/38 [01:53<02:14, 5.62s/it]\Examining click: 61%|████████████████████████████████████████████▊ | 23/38 [02:35<00:39, 2.62s/it]/Examining meson[version='>=0.64.0']: 68%|████████████████████████████████████▉ | 26/38 [02:37<00:22, 1.84s/it]\
Examining meson-python: 92%|█████████████████████████████████████████████████████████████▋ | 35/38 [03:26<00:10, 3.57s/it]/
Examining typing_extensions: 92%|█████████████████████████████████████████████████████████ | 35/38 [03:26<00:10, 3.57s/it]|Examining conflict for mpmath gmpy2: 5%|██▉ | 2/38 [00:14<04:22, 7.30s/it]\Examining conflict for mpmath openblas numpy pytest-xdist pytest asv matplotlib python cython pkg-config ninja click gmpy2 pythr\
Examining conflict for pytest pydevtool click: 24%|██████████▋ | 9/38 [01:33<05:12, 10.79s/it]\Examining conflict for numpy pythran libblas matplotlib: 42%|██████████████▎ | 16/38 [01:57<01:56, 5.30s/it]/
Examining conflict for sphinx pooch: 53%|████████████████████████████▍ | 20/38 [02:25<01:35, 5.29s/it]\Examining conflict for sphinx setuptools pooch pytest-xdist meson ipython cython pytest asv pytest-cov matplotlib: 68%|▋| 26/38/
Examining conflict for pytest-cov sphinx-design sphinx python setuptools cython-lint numpydoc pytest-xdist meson ipython cython /Examining conflict for sphinx setuptools pytest-xdist meson ipython cython pytest asv pytest-cov matplotlib: 89%|▉| 34/38 [02:5-
Examining conflict for pytest ipython: 95%|█████████████████████████████████████████████████▎ | 36/38 [03:07<00:04, 2.45s/it]/failed \ / /
UnsatisfiableError: The following specifications were found to be incompatible with each other:
Output in format: Requested package -> Available versions
Package appdirs conflicts for:
ipython -> black -> appdirs
pooch -> appdirs[version='>=1.3.0']
Package backports.functools_lru_cache conflicts for:
pytest -> wcwidth -> backports.functools_lru_cache
matplotlib -> backports.functools_lru_cache
Package colorama conflicts for:
meson-python -> colorama
numpydoc -> sphinx[version='>=1.8'] -> colorama[version='>=0.3.5|>=0.4.5']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> colorama[version='>=0.3.5|>=0.4.5']
ipython -> colorama
python -> pip -> colorama
click -> colorama
pytest -> colorama
pytest-cov -> pytest[version='>=4.6'] -> colorama
sphinx -> colorama[version='>=0.3.5|>=0.4.5']
pytest-timeout -> pytest[version='>=5.0.0'] -> colorama
rich-click -> click[version='>=7,<9'] -> colorama[version='>=0.4.0,<0.5.0']
sphinx-design -> sphinx[version='>=4,<6'] -> colorama[version='>=0.3.5|>=0.4.5']
pytest-xdist -> pytest[version='>=6.2.0'] -> colorama
Package vs2015_runtime conflicts for:
ipython -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
cython-lint -> cython[version='>=0.29.32'] -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0|>=14.16.27033|>=14.29.30139']
mypy -> psutil[version='>=4.0'] -> vs2015_runtime[version='>=14.15.26706|>=14.16.27012,<15.0a0|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
numpy -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
pooch -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
ruff -> vs2015_runtime[version='>=14.29.30139']
compilers -> c-compiler==1.2.0=h8ffe710_0 -> vs2015_runtime[version='>=14.16.27012']
doit[version='>=0.36.0'] -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
libblas[build=*openblas] -> libopenblas==0.3.21=pthreads_h02691f0_0 -> vs2015_runtime[version='>=14.16.27012|>=14.16.27033']
sphinx -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
numpy -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
rich-click -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pkg-config -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0']
typing_extensions -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
sphinx-design -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
setuptools[version='<60.0'] -> python[version='>=3.11,<3.12.0a0'] -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0|>=14.16.27033|>=14.29.30139']
pybind11 -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
meson-python -> ninja -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
asv[version='<0.5'] -> python[version='>=3.10,<3.11.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27012,<15.0a0|>=14.16.27033|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30139|>=14.29.30037|>=14.27.29016']
cython -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
ruff -> python[version='>=3.8,<3.9.0a0'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
pkg-config -> libglib[version='>=2.64.6,<3.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27033|>=14.29.30139|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
pybind11 -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
openblas -> vs2015_runtime[version='>=14.16.27012|>=14.16.27033|>=14.29.30139']
asv[version='<0.5'] -> vs2015_runtime[version='>=14.16.27012']
cython -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
numpydoc -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pytest-timeout -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pythran[version='>=0.11.0'] -> clang -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0|>=14.29.30139|>=14.16.27033']
ninja -> vc[version='>=14.1,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30139|>=14.29.30037']
gmpy2 -> python[version='>=3.11,<3.12.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27033|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
gmpy2 -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.29.30139']
pytest -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
python -> libsqlite[version='>=3.39.4,<4.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.29.30139|>=14.34.31931|>=14.32.31332|>=14.29.30037']
pydevtool -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
matplotlib -> vs2015_runtime[version='>=14.16.27012,<15.0a0']
openblas -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
ninja -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
threadpoolctl -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pydata-sphinx-theme==0.9.0 -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
mpmath -> python[version='>=3'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
pytest-xdist -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
mypy -> vs2015_runtime[version='>=14.16.27012|>=14.16.27033|>=14.29.30139']
pytest-cov -> coverage[version='>=5.2.1'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
click -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
python -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27012|>=14.16.27033|>=14.29.30139|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
Package python conflicts for:
cython -> python_abi=3.9[build=*_cp39] -> python[version='3.10.*|3.11.*|3.8.*|3.9.*|3.7.*|>=3.7|>=3.8']
ninja -> python_abi=3.8[build=*_cp38] -> python[version='3.6.*|3.7.*|3.8.*']
pytest-xdist -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|>=3.7|3.4.*|>=2.7,<2.8.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
gmpy2 -> python_abi=3.11[build=*_cp311] -> python[version='3.10.*|3.11.*|3.9.*|3.8.*|3.7.*']
pooch -> appdirs[version='>=1.3.0'] -> python[version='2.7.*|3.5.*|3.6.*|3.4.*|>=3.7|>=2.7|>=3.9,<3.10.0a0|>=3.10,<3.11.0a0|>=3.7,<4.0|>=3.6,<4.0|>=3.11,<3.12.0a0']
pooch -> python[version='>=2.7,<2.8.0a0|>=3.5|>=3.6|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
pytest -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.8|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
click -> colorama -> python[version='3.10.*|3.8.*|>=3.7|3.11.*|3.9.*|3.7.*']
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2'] -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.7|>=3.8|>=3.11,<3.12.0a0|3.4.*']
pytest -> attrs[version='>=19.2.0'] -> python[version='2.7|>=3.4|3.10.*|3.8.*|3.9.*|>=2.7|>=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*|>=3.5|>=3.7|>=3.6|3.11.*|>=3|3.7.*|>=3.4|>=3.8.0a,<3.9.0a0']
python
pytest-cov -> coverage[version='>=5.2.1'] -> python[version='>=2.7|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.8|>=3.7']
numpydoc -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.7|3.4.*|>=3.7,<3.8.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.10,<3.11.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=2.7,<2.8.0a0']
rich-click -> click[version='>=7,<9'] -> python[version='>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.7.0|>=3.6.2']
ninja -> python[version='>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
sphinx-design -> sphinx[version='>=4,<6'] -> python[version='>=3.10,<3.11.0a0|>=3.7|>=3.7,<3.8.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.5']
mypy -> python[version='3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5|>=3.4|>=3.5,<3.6.0a0|>=3.6']
meson[version='>=0.64.0'] -> python[version='>=3.10,<3.11.0a0|>=3.5.2|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0']
numpy -> python[version='3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
cython-lint -> python[version='>=3.6|>=3.7']
pytest-cov -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|3.4.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.7,<3.8.0a0|>=3.5,<3.6.0a0']
doit[version='>=0.36.0'] -> python[version='>=3.8']
pythran[version='>=0.11.0'] -> python[version='>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0']
asv[version='<0.5'] -> python_abi=3.10[build=*_cp310] -> python[version='3.10.*|>=3.7|>=3.8|>=3.11,<3.12.0a0|3.9.*|3.7.*|3.8.*']
meson-python -> python[version='>=3.10,<3.11.0a0|>=3.7|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0']
ipython -> python[version='2.7.*|3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.8|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=2.7,<2.8.0a0|3.4.*']
sphinx -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.5|>=3.6|>=3.7|>=3.8|>=3.6,<3.7.0a0|>=3.7,<3.8.0a0|>=3.5,<3.6.0a0|3.4.*|>=3.9,<3.10.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0']
matplotlib -> python[version='3.4.*|3.5.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
cython -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
matplotlib -> tornado[version='>=5'] -> python[version='2.7.*|3.6.*|3.10.*|3.8.*|3.9.*|3.11.*|3.7.*|>=3.6|>=3.0,<=3.9|>=3|>=3.7|>=3.8|<3']
pydevtool -> python[version='>=3.8']
pydata-sphinx-theme==0.9.0 -> python[version='>=3.7']
sphinx -> alabaster[version='>=0.7,<0.8'] -> python[version='>=2.7|>=2|>=3.11,<3.12.0a0|>=3.4|>=3.7,<4.0|>=3.6,<4.0|>=3']
numpy -> python_abi=3.10[build=*_cp310] -> python[version='2.7.*|3.10.*|3.11.*|3.9.*|3.8.*|3.7.*']
threadpoolctl -> python[version='>=3.5|>=3.6']
mypy -> mypy_extensions[version='>=0.4.3'] -> python[version='2.7.*|>=2.7,<2.8.0a0|3.4.*|3.8.*|>=3.7|3.10.*|3.9.*|3.11.*|3.7.*|>=2.7|>=3']
typing_extensions -> python[version='>=2.7,<2.8.0a0|>=3.5|>=3.6|>=3.7|>=3.6,<3.7.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0|>=3.5,<3.6.0a0|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0']
ruff -> python_abi=3.8[build=*_cp38] -> python[version='3.10.*|3.8.*|3.11.*|3.9.*']
pytest-xdist -> execnet[version='>=1.1'] -> python[version='2.7|>=3.5|>=3.8|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=2.7']
setuptools[version='<60.0'] -> python_abi=3.11[build=*_cp311] -> python[version='3.10.*|3.11.*|3.8.*|3.7.*|3.9.*|>=3.7|>=2.7']
ruff -> python[version='>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0']
pydevtool -> doit -> python[version='3.5.*|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.6.*|>=2.7,<2.8.0a0']
pythran[version='>=0.11.0'] -> beniget=0.4 -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|3.4.*|>=3.6,<3.7.0a0|>=2.7,<2.8.0a0|>=3.5,<3.6.0a0|>=3.4|3.11.*|3.8.*|3.10.*|3.9.*|3.7.*']
pybind11 -> python_abi=3.10[build=*_cp310] -> python[version='3.10.*|3.8.*|3.11.*|3.9.*|3.7.*']
asv[version='<0.5'] -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
sphinx-design -> python[version='>=3.6']
gmpy2 -> python[version='3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
pytest-timeout -> python[version='2.7.*|3.4.*|3.5.*|3.6.*|>=3.6|>=3.8,<3.9.0a0|>=2.7,<2.8.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
doit[version='>=0.36.0'] -> cloudpickle -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|3.4.*|>=3|>=2.7,<2.8.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.10,<3.11.0a0']
pybind11 -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
meson-python -> colorama -> python[version='!=3.0,!=3.1,!=3.2,!=3.3,!=3.4|2.7.*|3.5.*|3.6.*|>=3.6|3.4.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.5.2|2.7.*|>=3.6']
ipython -> backcall -> python[version='3.10.*|3.9.*|>=2.7|>=3|>=3.5|>=3.6|>=3.7|3.11.*|3.8.*|>=3.11,<3.12.0a0|3.7.*']
click -> python[version='2.7.*|3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.8|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|3.4.*|>=3|>=3.6|>=2.7,<2.8.0a0|>=3.5,<3.6.0a0']
typing_extensions -> python_abi=3.6[build=*_cp36m] -> python[version='2.7.*|3.6.*|3.7.*|3.8.*|>=3|3.5.*']
pytest-timeout -> pytest[version='>=5.0.0'] -> python[version='>=3.10,<3.11.0a0|>=3.8|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0']
cython-lint -> cython[version='>=0.29.32'] -> python[version='2.7.*|2.7.*|>=3.5|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|3.6.*|3.5.*|3.4.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.6.1']
rich-click -> python[version='>=3.6|>=3.7']
numpydoc -> jinja2[version='>=2.10'] -> python[version='>=3.6|>=3.8']
setuptools[version='<60.0'] -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
pkg-config -> glib[version='>=2.58.3,<3.0a0'] -> python[version='*|>=2.7|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0']
pydata-sphinx-theme==0.9.0 -> beautifulsoup4 -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.6|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0|>=3.5,<3.6.0a0|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|3.4.*|>=2.7|>=3.8|>=3.5']
Package six conflicts for:
ipython -> prompt-toolkit[version='>3.0.1,<3.1.0'] -> six[version='>=1.13.0|>=1.9.0']
asv[version='<0.5'] -> six
pydevtool -> doit -> six
pytest -> six[version='>=1.10.0']
sphinx -> six[version='>=1.5']
pydata-sphinx-theme==0.9.0 -> packaging -> six[version='>=1.5']
pooch -> packaging[version='>=20.0'] -> six
numpy -> mkl-service[version='>=2.3.0,<3.0a0'] -> six
pytest-xdist -> six
pytest-timeout -> pytest[version='>=3.6.0'] -> six[version='>=1.10.0']
numpydoc -> sphinx[version='>=1.8'] -> six[version='>=1.5']
pytest -> more-itertools[version='>=4.0.0'] -> six[version='>=1.0.0,<2.0.0|>=1.13.0']
pytest-xdist -> pytest[version='>=4.4.0'] -> six[version='>=1.10.0']
pytest-cov -> pytest[version='>=4.6'] -> six[version='>=1.10.0']
matplotlib -> cycler -> six[version='>=1.5']
Package typing_extensions conflicts for:
ipython -> black -> typing_extensions[version='>=3.10.0.0|>=3.10|>=3.10,!=3.10.0.1|>=3.7.4']
typing_extensions
click -> importlib-metadata -> typing_extensions[version='>=3.6.4']
rich-click -> rich[version='>=10'] -> typing_extensions[version='>=3.7.4,<4.0.0|>=3.7.4,<5.0.0|>=4.0.0,<5.0.0']
pytest -> importlib-metadata[version='>=0.12'] -> typing_extensions[version='>=3.6.4']
matplotlib -> kiwisolver -> typing_extensions
sphinx -> importlib-metadata[version='>=4.8'] -> typing_extensions[version='>=3.6.4']
mypy -> typing_extensions[version='>=3.10|>=3.7.4']
meson-python -> typing-extensions[version='>=3.7.4'] -> typing_extensions[version='3.10.0.0|3.10.0.0|3.10.0.1|3.10.0.2|4.0.0|4.0.1|4.1.1|4.2.0|4.2.0|4.3.0|4.4.0|3.7.4.3|3.7.4.2|3.7.4.1|3.7.4.1|3.7.4.1|3.7.4.1|3.7.4|4.4.0|4.4.0|4.4.0|4.4.0|4.3.0|4.3.0|4.3.0|4.3.0|4.1.1|3.10.0.2|3.7.4.3|3.7.4.1|3.7.4',build='py36_0|pyh06a4308_0|py38haa95532_0|py310haa95532_0|py37haa95532_0|py39haa95532_0|py38haa95532_0|py37haa95532_0|py37_0|py36_0|py36h9f0ad1d_1|py36h9f0ad1d_2|py_0|py_0|pyha770c72_0|pyha770c72_0|pyha770c72_0|pyha770c72_0|pyha770c72_0|pyha770c72_1|pyha770c72_0|pyha770c72_0|pyha770c72_0|py38h32f6830_3|py310haa95532_0|py39haa95532_0|pyh06a4308_0|pyh06a4308_0|pyh06a4308_0|py38_0']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4'] -> typing_extensions[version='>=3.6.4']
Package pypy3.7 conflicts for:
pytest -> pypy3.7[version='>=7.3.5|>=7.3.7']
ipython -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
pythran[version='>=0.11.0'] -> pypy3.7[version='>=7.3.7']
click -> pypy3.7[version='>=7.3.5|>=7.3.7']
pybind11 -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
asv[version='<0.5'] -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
sphinx -> docutils[version='>=0.18,<0.20'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
numpydoc -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pytest-timeout -> pytest[version='>=5.0.0'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
meson-python -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pydata-sphinx-theme==0.9.0 -> docutils!=0.17.0 -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
threadpoolctl -> python[version='>=3.6'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pybind11 -> pypy3.7[version='>=7.3.5|>=7.3.7']
pytest-xdist -> pytest[version='>=6.2.0'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
cython -> pypy3.7[version='>=7.3.5|>=7.3.7']
pythran[version='>=0.11.0'] -> numpy[version='>=1.20.3,<2.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*|>=7.3.5']
cython -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
mypy -> mypy_extensions[version='>=0.4.3'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
asv[version='<0.5'] -> pypy3.7[version='>=7.3.5|>=7.3.7']
pooch -> python[version='>=3.6'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.7']
cython-lint -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
python -> pypy3.7[version='7.3.5.*|7.3.7.*']
gmpy2 -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
python -> python_abi==3.7[build=*_pypy37_pp73] -> pypy3.7=7.3
numpy -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
ninja -> python -> pypy3.7[version='7.3.5.*|7.3.7.*']
pytest -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
setuptools[version='<60.0'] -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
numpy -> pypy3.7[version='>=7.3.5|>=7.3.7']
matplotlib -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4'] -> pypy3.7[version='>=7.3.5|>=7.3.7']
meson[version='>=0.64.0'] -> python[version='>=3.5.2'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.7|>=7.3.5']
mpmath -> python[version='>=3'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
ipython -> pypy3.7[version='>=7.3.5|>=7.3.7']
typing_extensions -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
setuptools[version='<60.0'] -> pypy3.7[version='>=7.3.5|>=7.3.7']
matplotlib -> pypy3.7[version='>=7.3.5|>=7.3.7']
sphinx-design -> python[version='>=3.6'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pytest-cov -> coverage[version='>=5.2.1'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
click -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
rich-click -> click[version='>=7,<9'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
Package numpy conflicts for:
matplotlib -> numpy[version='1.10.*|1.11.*|>=1.14.6,<2.0a0']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> numpy[version='>=1.11.3,<2.0a0|>=1.15.4,<2.0a0|>=1.16.5,<2.0a0|>=1.16.6,<2.0a0|>=1.17|>=1.19.2,<2.0a0|>=1.19|>=1.20.3,<2.0a0|>=1.21.6,<2.0a0|>=1.23.5,<2.0a0|>=1.23.4,<2.0a0|>=1.21.5,<2.0a0|>=1.19.5,<2.0a0|>=1.18.5,<2.0a0|>=1.21.4,<2.0a0|>=1.21.2,<2.0a0|>=1.17.5,<2.0a0|>=1.19.4,<2.0a0']
numpy
pythran[version='>=0.11.0'] -> numpy[version='>=1.16.6,<2.0a0|>=1.18.5,<2.0a0|>=1.19.5,<2.0a0|>=1.20.3,<2.0a0|>=1.23.5,<2.0a0|>=1.21.6,<2.0a0|>=1.23.4,<2.0a0|>=1.21.5,<2.0a0|>=1.21.4,<2.0a0']
Package tomli conflicts for:
pytest -> tomli[version='>=1.0.0']
pytest-cov -> coverage[version='>=5.2.1'] -> tomli[version='>=1.0.0']
mypy -> tomli[version='>=1.1.0']
meson-python -> tomli[version='>=1.0.0']
pytest-timeout -> pytest[version='>=5.0.0'] -> tomli[version='>=1.0.0']
pytest-xdist -> pytest[version='>=6.2.0'] -> tomli[version='>=1.0.0']
ipython -> black -> tomli[version='>=0.2.6,<2.0.0|>=1.1.0']
Package setuptools conflicts for:
pytest -> setuptools[version='>=40.0']
meson[version='>=0.64.0'] -> setuptools
pytest-xdist -> pytest[version='>=4.4.0'] -> setuptools[version='>=40.0']
python -> pip -> setuptools
numpydoc -> jinja2[version='>=2.10'] -> setuptools
sphinx -> setuptools
matplotlib -> setuptools
ipython -> pygments[version='>=2.4.0'] -> setuptools
cython -> setuptools
pytest-cov -> setuptools
cython-lint -> pycodestyle -> setuptools
sphinx-design -> sphinx[version='>=4,<6'] -> setuptools
meson-python -> meson[version='>=0.63.3'] -> setuptools
setuptools[version='<60.0']
pytest-xdist -> setuptools
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> setuptools
pytest-cov -> pytest[version='>=4.6'] -> setuptools[version='>=40.0']
asv[version='<0.5'] -> setuptools
ipython -> setuptools[version='>=18.5']
pytest-timeout -> pytest[version='>=5.0.0'] -> setuptools[version='>=40.0']
Package pathlib2 conflicts for:
pytest-xdist -> pytest[version='>=4.4.0'] -> pathlib2[version='>=2.2.0']
pytest -> pathlib2[version='>=2.2.0']
click -> importlib-metadata -> pathlib2
ipython -> pathlib2
pytest-cov -> pytest[version='>=4.6'] -> pathlib2[version='>=2.2.0']
pytest-timeout -> pytest[version='>=5.0.0'] -> pathlib2[version='>=2.2.0']
Package vs2008_runtime conflicts for:
numpy -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
click -> python -> vs2008_runtime
matplotlib -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2'] -> vs2008_runtime
pytest-cov -> python -> vs2008_runtime
numpydoc -> python -> vs2008_runtime
pooch -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
asv[version='<0.5'] -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
python -> vc[version='>=9,<10.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
ninja -> vc=9 -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
meson-python -> ninja -> vs2008_runtime
ninja -> vs2008_runtime
pytest-timeout -> python -> vs2008_runtime
setuptools[version='<60.0'] -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
sphinx -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
pytest-xdist -> python -> vs2008_runtime
mpmath -> python -> vs2008_runtime
cython -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
pybind11 -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
python -> vs2008_runtime
pytest -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
ipython -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
typing_extensions -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
Package typed-ast conflicts for:
ipython -> black -> typed-ast[version='>=1.4.0|>=1.4.2']
mypy -> typed-ast[version='>=1.1.0,<1.2.0|>=1.2.0,<1.3.0|>=1.3.1,<1.4.0|>=1.4.0,<1.5.0|>=1.4.0,<2']
Package wincertstore conflicts for:
sphinx -> setuptools -> wincertstore[version='>=0.2']
matplotlib -> setuptools -> wincertstore[version='>=0.2']
meson[version='>=0.64.0'] -> setuptools -> wincertstore[version='>=0.2']
pytest-cov -> setuptools -> wincertstore[version='>=0.2']
pytest-xdist -> setuptools -> wincertstore[version='>=0.2']
asv[version='<0.5'] -> setuptools -> wincertstore[version='>=0.2']
ipython -> setuptools[version='>=18.5'] -> wincertstore[version='>=0.2']
setuptools[version='<60.0'] -> wincertstore[version='>=0.2']
pytest -> setuptools[version='>=40.0'] -> wincertstore[version='>=0.2']
cython -> setuptools -> wincertstore[version='>=0.2']
Package importlib-metadata conflicts for:
sphinx-design -> sphinx[version='>=4,<6'] -> importlib-metadata[version='>=4.4|>=4.8']
rich-click -> click[version='>=7,<9'] -> importlib-metadata
pytest-timeout -> pytest[version='>=5.0.0'] -> importlib-metadata[version='>=0.12']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> importlib-metadata[version='>=4.4|>=4.8']
pytest-cov -> pytest[version='>=4.6'] -> importlib-metadata[version='>=0.12']
pytest -> importlib_metadata[version='>=0.12'] -> importlib-metadata[version='>=1.1.3,<1.1.4.0a0|>=1.5.0,<1.5.1.0a0|>=1.5.2,<1.5.3.0a0|>=1.6.0,<1.6.1.0a0|>=1.6.1,<1.6.2.0a0|>=1.7.0,<1.7.1.0a0|>=2.0.0,<2.0.1.0a0|>=3.0.0,<3.0.1.0a0|>=3.1.0,<3.1.1.0a0|>=3.1.1,<3.1.2.0a0|>=3.10.0,<3.10.1.0a0|>=3.10.1,<3.10.2.0a0|>=4.0.1,<4.0.2.0a0|>=4.10.0,<4.10.1.0a0|>=4.10.1,<4.10.2.0a0|>=4.11.0,<4.11.1.0a0|>=4.11.1,<4.11.2.0a0|>=4.11.2,<4.11.3.0a0|>=4.11.3,<4.11.4.0a0|>=4.11.4,<4.11.5.0a0|>=5.0.0,<5.0.1.0a0|>=5.1.0,<5.1.1.0a0|>=5.2.0,<5.2.1.0a0|>=6.0.0,<6.0.1.0a0|>=4.9.0,<4.9.1.0a0|>=4.8.3,<4.8.4.0a0|>=4.8.2,<4.8.3.0a0|>=4.8.1,<4.8.2.0a0|>=4.8.0,<4.8.1.0a0|>=4.7.1,<4.7.2.0a0|>=4.7.0,<4.7.1.0a0|>=4.6.4,<4.6.5.0a0|>=4.6.3,<4.6.4.0a0|>=4.6.2,<4.6.3.0a0|>=4.6.1,<4.6.2.0a0|>=4.6.0,<4.6.1.0a0|>=4.5.0,<4.5.1.0a0|>=4.4.0,<4.4.1.0a0|>=4.3.1,<4.3.2.0a0|>=4.3.0,<4.3.1.0a0|>=4.2.0,<4.2.1.0a0|>=3.9.1,<3.9.2.0a0|>=3.9.0,<3.9.1.0a0|>=3.8.1,<3.8.2.0a0|>=3.8.0,<3.8.1.0a0|>=3.7.3,<3.7.4.0a0|>=3.7.2,<3.7.3.0a0|>=3.7.0,<3.7.1.0a0|>=3.6.0,<3.6.1.0a0|>=3.4.0,<3.4.1.0a0|>=3.3.0,<3.3.1.0a0']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4']
pydevtool -> doit -> importlib-metadata[version='>=4.4']
numpydoc -> sphinx[version='>=1.8'] -> importlib-metadata[version='>=4.4|>=4.8']
pytest -> importlib-metadata[version='>=0.12']
pytest-xdist -> pytest[version='>=6.2.0'] -> importlib-metadata[version='>=0.12']
sphinx -> importlib-metadata[version='>=4.4|>=4.8']
click -> importlib-metadata
Package vs2010_runtime conflicts for:
numpydoc -> python -> vs2010_runtime
matplotlib -> python=3.4 -> vs2010_runtime
pybind11 -> python=3.4 -> vs2010_runtime
ipython -> python=3.4 -> vs2010_runtime
pytest-timeout -> python -> vs2010_runtime
pytest-xdist -> python -> vs2010_runtime
asv[version='<0.5'] -> python=3.4 -> vs2010_runtime
sphinx -> python=3.4 -> vs2010_runtime
setuptools[version='<60.0'] -> python=3.4 -> vs2010_runtime
mpmath -> python[version='>=3'] -> vs2010_runtime
pytest -> python=3.4 -> vs2010_runtime
ninja -> python -> vs2010_runtime
mypy -> python[version='>=3.4'] -> vs2010_runtime
click -> python -> vs2010_runtime
pytest-cov -> python -> vs2010_runtime
python -> vs2010_runtime
cython -> python=3.4 -> vs2010_runtime
Package packaging conflicts for:
pytest-cov -> pytest[version='>=4.6'] -> packaging
meson-python -> pyproject-metadata[version='>=0.6.1'] -> packaging[version='>=19.0']
ipython -> black -> packaging[version='>=22.0']
sphinx -> packaging[version='>=21.0']
pooch -> packaging[version='>=20.0']
pytest -> packaging
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> packaging[version='>=20.0']
python -> pip -> packaging
sphinx-design -> sphinx[version='>=4,<6'] -> packaging[version='>=21.0']
pytest-xdist -> pytest[version='>=6.2.0'] -> packaging
pydata-sphinx-theme==0.9.0 -> packaging
pytest-timeout -> pytest[version='>=5.0.0'] -> packaging
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> packaging[version='>=21.0']
numpydoc -> sphinx[version='>=1.8'] -> packaging[version='>=21.0']
Package zipp conflicts for:
click -> importlib-metadata -> zipp[version='>=0.5']
sphinx -> importlib-metadata[version='>=4.8'] -> zipp[version='>=0.5']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4'] -> zipp[version='>=0.5']
pytest -> importlib-metadata[version='>=0.12'] -> zipp[version='>=0.5']
Package sqlite conflicts for:
gmpy2 -> python[version='>=3.11,<3.12.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
numpydoc -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
ruff -> python[version='>=3.8,<3.9.0a0'] -> sqlite[version='>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.39.2,<4.0a0|>=3.39.1,<4.0a0|>=3.38.2,<4.0a0']
doit[version='>=0.36.0'] -> python[version='>=3.8'] -> sqlite[version='>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0']
threadpoolctl -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
cython -> python[version='>=3.9,<3.10.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.39.1,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
cython-lint -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
matplotlib -> pypy3.8[version='>=7.3.9'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.2,<4.0a0|>=3.39.1,<4.0a0|>=3.39.2,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
setuptools[version='<60.0'] -> python[version='>=3.11,<3.12.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.2,<4.0a0|>=3.39.1,<4.0a0|>=3.39.2,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
python -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest-timeout -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
click -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.39.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
meson-python -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0']
pooch -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pybind11 -> python[version='>=3.10,<3.11.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.39.1,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pydata-sphinx-theme==0.9.0 -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0']
mypy -> python[version='>=3.8,<3.9.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pythran[version='>=0.11.0'] -> pypy3.8[version='>=7.3.9'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.2,<4.0a0|>=3.39.1,<4.0a0|>=3.39.2,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0']
pydevtool -> python[version='>=3.8'] -> sqlite[version='>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0']
ninja -> python -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
ipython -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest-cov -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
meson[version='>=0.64.0'] -> python[version='>=3.5.2'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
asv[version='<0.5'] -> python[version='>=3.10,<3.11.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.30.0,<4.0a0|>=3.35.1,<4.0a0']
mpmath -> python[version='>=3'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest-xdist -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
sphinx -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
typing_extensions -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.39.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
rich-click -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
sphinx-design -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
numpy -> python[version='>=3.10,<3.11.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.39.1,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
python -> pypy3.9=7.3.11 -> sqlite[version='>=3.39.1,<4.0a0']
Package certifi conflicts for:
pytest-cov -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
sphinx -> requests[version='>=2.25.0'] -> certifi[version='>=2016.09|>=2016.9.26|>=2017.4.17']
asv[version='<0.5'] -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
meson[version='>=0.64.0'] -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
setuptools[version='<60.0'] -> certifi[version='>=2016.09|>=2016.9.26']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> certifi[version='>=2016.09|>=2016.9.26|>=2020.06.20']
cython -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
pytest -> setuptools[version='>=40.0'] -> certifi[version='>=2016.09|>=2016.9.26']
pytest-xdist -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
ipython -> setuptools[version='>=18.5'] -> certifi[version='>=2016.09|>=2016.9.26']
pooch -> requests[version='>=2.19.0'] -> certifi[version='>=2017.4.17']
Package configparser conflicts for:
pytest -> importlib-metadata[version='>=0.12'] -> configparser[version='>=3.5']
pydevtool -> doit -> configparser
click -> importlib-metadata -> configparser[version='>=3.5']
Package libflang conflicts for:
compilers -> fortran-compiler==1.2.0=h9655429_0 -> libflang[version='<6.0.0.a0|>=5.0.0|>=5.0.0,<6.0.0.a0']
libblas[build=*openblas] -> libopenblas==0.3.21=pthreads_h02691f0_0 -> libflang[version='<6.0.0.a0|>=5.0.0|>=5.0.0,<6.0.0.a0']
numpy -> libflang[version='<6.0.0.a0|>=5.0.0']
matplotlib -> numpy[version='>=1.14.6,<2.0a0'] -> libflang[version='<6.0.0.a0|>=5.0.0']
openblas -> libflang[version='<6.0.0.a0|>=5.0.0|>=5.0.0,<6.0.0.a0']
numpy -> openblas[version='>=0.2.20,<0.2.21.0a0'] -> libflang
Package sphinx conflicts for:
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2']
sphinx
numpydoc -> sphinx[version='>=1.8|>=4.2|>=3.0']
sphinx-design -> sphinx[version='>=3,<5|>=4,<6']
Package libopenblas conflicts for:
numpy -> libblas[version='>=3.9.0,<4.0a0'] -> libopenblas[version='0.3.10|0.3.12|0.3.12|0.3.15|0.3.17|0.3.18|0.3.20|0.3.21|0.3.9|0.3.8|0.3.7|0.3.7|0.3.6',build='hcfe7411_0|pthreads_h1662909_0|pthreads_h02691f0_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_h1662909_1|hc8a65f3_0|hcfe7411_0|h29e5d5d_1|h29e5d5d_0|h29e5d5d_6']
libblas[build=*openblas] -> openblas[version='>=0.3.6,<0.3.7.0a0'] -> libopenblas==0.3.6[build='h29e5d5d_3|h29e5d5d_4|h29e5d5d_5']
libblas[build=*openblas] -> libopenblas[version='0.3.10|0.3.12|0.3.12|0.3.15|0.3.17|0.3.18|0.3.20|0.3.21|0.3.9|0.3.8|0.3.7|0.3.7|0.3.6',build='hcfe7411_0|pthreads_h1662909_0|pthreads_h02691f0_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_h1662909_1|hc8a65f3_0|hcfe7411_0|h29e5d5d_1|h29e5d5d_0|h29e5d5d_6']
openblas -> libopenblas[version='0.3.10|0.3.10|0.3.10|0.3.10|0.3.10|0.3.10|0.3.12|0.3.12|0.3.13|0.3.15|0.3.15|0.3.16|0.3.17|0.3.17|0.3.18|0.3.20|0.3.20|0.3.21|0.3.9|0.3.8|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.6|0.3.6|0.3.6|0.3.6',build='h29e5d5d_3|h29e5d5d_4|h29e5d5d_5|hcfe7411_4|hcfe7411_7|hcfe7411_0|pthreads_h1662909_4|pthreads_hc469a61_1|pthreads_h02691f0_2|pthreads_hc140b1d_3|pthreads_h02691f0_1|pthreads_h02691f0_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_h1662909_1|pthreads_h1662909_0|pthreads_h1662909_5|pthreads_h1662909_3|pthreads_h1662909_2|pthreads_h1662909_1|hc8a65f3_0|hcfe7411_0|hcfe7411_6|hcfe7411_5|h29e5d5d_3|h29e5d5d_2|h29e5d5d_1|h29e5d5d_0|h29e5d5d_6']
Package ca-certificates conflicts for:
matplotlib -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
pytest-timeout -> python -> ca-certificates
asv[version='<0.5'] -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
python -> ca-certificates
sphinx -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
setuptools[version='<60.0'] -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
numpy -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
mpmath -> python -> ca-certificates
pytest -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
ninja -> python -> ca-certificates
ipython -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
cython -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
pybind11 -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
numpydoc -> python -> ca-certificates
pooch -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
typing_extensions -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
pytest-xdist -> python -> ca-certificates
click -> python -> ca-certificates
pytest-cov -> python -> ca-certificates
Package pytest conflicts for:
pytest
pytest-xdist -> pytest[version='>=2.4.2|>=2.7.0|>=3.0.0|>=4.4.0|>=6.0.0|>=6.2.0']
pytest-xdist -> pytest-forked -> pytest[version='>=2.6.0|>=3.1.0|>=3.10']
pytest-cov -> pytest[version='>=3.6|>=4.6']
pytest-timeout -> pytest[version='>=2.8.0|>=3.6.0|>=5.0.0']
Package numpydoc conflicts for:
ipython -> jedi[version='>=0.10,<0.18'] -> numpydoc
numpydoc
Package expat conflicts for:
pythran[version='>=0.11.0'] -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
pytest -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
matplotlib -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
cython -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
numpy -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0|>=2.4.7,<3.0a0']
ipython -> pypy3.7[version='>=7.3.7'] -> expat[version='>=2.4.1,<3.0a0']
click -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
python -> pypy3.9=7.3.11 -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
setuptools[version='<60.0'] -> pypy3.8[version='>=7.3.8'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
ruff -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
asv[version='<0.5'] -> pypy3.7[version='>=7.3.7'] -> expat[version='>=2.4.1,<3.0a0']
pybind11 -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
Package typing conflicts for:
sphinx -> typing
pydata-sphinx-theme==0.9.0 -> sphinx -> typing
mypy -> typing[version='>=3.5.3']
typing_extensions -> typing[version='>=3.6.2|>=3.7.4']
mypy -> typing_extensions[version='>=3.7.4'] -> typing[version='>=3.7.4']
numpydoc -> sphinx[version='>=1.8'] -> typing
Package typing-extensions conflicts for:
meson-python -> typing-extensions[version='>=3.7.4']
matplotlib -> kiwisolver -> typing-extensions
Package ucrt conflicts for:
pytest-cov -> coverage[version='>=5.2.1'] -> ucrt[version='>=10.0.20348.0']
openblas -> vs2015_runtime[version='>=14.29.30139'] -> ucrt[version='>=10.0.20348.0']
gmpy2 -> ucrt[version='>=10.0.20348.0']
cython-lint -> cython[version='>=0.29.32'] -> ucrt[version='>=10.0.20348.0']
click -> pypy3.8[version='>=7.3.9'] -> ucrt[version='>=10.0.20348.0']
setuptools[version='<60.0'] -> pypy3.8[version='>=7.3.8'] -> ucrt[version='>=10.0.20348.0']
pytest -> pypy3.8[version='>=7.3.9'] -> ucrt[version='>=10.0.20348.0']
ruff -> ucrt[version='>=10.0.20348.0']
pytest-xdist -> psutil[version='>=3.0'] -> ucrt[version='>=10.0.20348.0']
pkg-config -> libglib[version='>=2.64.6,<3.0a0'] -> ucrt[version='>=10.0.20348.0']
mypy -> ucrt[version='>=10.0.20348.0']
mpmath -> gmpy2 -> ucrt[version='>=10.0.20348.0']
pythran[version='>=0.11.0'] -> numpy[version='>=1.20.3,<2.0a0'] -> ucrt[version='>=10.0.20348.0']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> ucrt[version='>=10.0.20348.0']
pybind11 -> ucrt[version='>=10.0.20348.0']
ninja -> vs2015_runtime[version='>=14.16.27033'] -> ucrt[version='>=10.0.20348.0']
cython -> ucrt[version='>=10.0.20348.0']
python -> libsqlite[version='>=3.39.4,<4.0a0'] -> ucrt[version='>=10.0.20348.0']
numpy -> ucrt[version='>=10.0.20348.0']
asv[version='<0.5'] -> vs2015_runtime[version='>=14.16.27012'] -> ucrt[version='>=10.0.20348.0']
Package dataclasses conflicts for:
ipython -> black -> dataclasses[version='>=0.6']
rich-click -> rich[version='>=10'] -> dataclasses[version='>=0.7,<0.9']
Package flit-core conflicts for:
typing_extensions -> flit-core[version='>=3.6,<4']
mypy -> typing_extensions[version='>=3.10'] -> flit-core[version='>=3.6,<4']
Package doit conflicts for:
doit[version='>=0.36.0']
pydevtool -> doit
Package pytz conflicts for:
sphinx -> babel[version='>=2.9'] -> pytz[version='>=2015.7']
matplotlib -> pytz
Package requests conflicts for:
python -> pip -> requests
sphinx -> requests[version='>2.0.0|>=2.0.0|>=2.25.0|>=2.5.0']
sphinx-design -> sphinx[version='>=4,<6'] -> requests[version='>=2.5.0']
pooch -> requests[version='>=2.19.0']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> requests[version='>2.0.0|>=2.0.0|>=2.25.0|>=2.5.0']
numpydoc -> sphinx[version='>=1.8'] -> requests[version='>2.0.0|>=2.0.0|>=2.25.0|>=2.5.0']
Package libblas conflicts for:
numpy -> libblas[version='>=3.8.0,<4.0a0|>=3.9.0,<4.0a0']
pythran[version='>=0.11.0'] -> numpy[version='>=1.20.3,<2.0a0'] -> libblas[version='>=3.8.0,<4.0a0|>=3.9.0,<4.0a0']
numpy -> libcblas[version='>=3.9.0,<4.0a0'] -> libblas[version='3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0.*|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0.*',build='5_openblas|5_blis|6_blis|6_openblas|7_blis|8_blis|8_openblas|8_mkl|9_blis|10_openblas|11_mkl|11_openblas|12_blis|13_mkl|14_openblas|15_blis|16_openblas|16_mkl|17_openblas|18_mkl|0_blis|0_openblas|1_openblas|2_openblas|4_blis|5_blis|5_openblas|5_mkl|6_blis|8_openblas|8_mkl|9_blis|9_openblas|10_openblas|11_win64_blis|13_win64_openblas|13_win64_mkl|14_win64_mkl|16_win64_openblas|16_win64_blis|16_win64_mkl|15_win64_blis|15_win64_mkl|15_win64_openblas|14_win64_openblas|14_win64_blis|13_win64_blis|12_win64_mkl|12_win64_openblas|12_win64_blis|11_win64_mkl|11_win64_openblas|10_mkl|10_blis|9_mkl|8_blis|7_mkl|7_openblas|7_blis|6_mkl|6_openblas|4_mkl|4_openblas|3_openblas|3_blis|2_blis|1_blis|21_mkl|20_mkl|19_mkl|16_blis|15_openblas|15_mkl|14_mkl|14_blis|13_openblas|13_blis|12_mkl|12_openblas|11_blis|10_mkl|10_blis|9_mkl|9_openblas|7_openblas|7_mkl|6_mkl|5_mkl|4_mkl|4_openblas|4_blis']
matplotlib -> numpy=1.11 -> libblas[version='>=3.8.0,<4.0a0|>=3.9.0,<4.0a0']
libblas[build=*openblas]
Package pygments conflicts for:
sphinx -> pygments[version='>2.0|>=2.0|>=2.12|>=2.13']
rich-click -> rich[version='>=10'] -> pygments[version='>=2.14.0,<3.0.0|>=2.6.0,<3.0.0']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> pygments[version='>2.0|>=2.0|>=2.12|>=2.13']
ipython -> pygments[version='>=2.4.0']
numpydoc -> sphinx[version='>=1.8'] -> pygments[version='>2.0|>=2.0|>=2.12|>=2.13']
sphinx-design -> sphinx[version='>=4,<6'] -> pygments[version='>=2.0|>=2.12']
Package pyparsing conflicts for:
pooch -> packaging[version='>=20.0'] -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
matplotlib -> pyparsing
pydata-sphinx-theme==0.9.0 -> packaging -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
sphinx -> packaging[version='>=21.0'] -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
pytest -> packaging -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> pyparsing[version='>=2.0.3,!=2.0.4,!=2.1.2,!=2.1.6|>=2.2.1|>=2.3.1']
Package ninja conflicts for:
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2']
meson-python -> ninja
meson-python -> meson[version='>=0.63.3'] -> ninja[version='>=1.8.2']
ninja
Package decorator conflicts for:
ipython -> decorator
pythran[version='>=0.11.0'] -> decorator
Package llvm-openmp conflicts for:
openblas -> libflang[version='>=5.0.0'] -> llvm-openmp[version='11.0.1|>=11.0.1']
numpy -> libflang[version='>=5.0.0'] -> llvm-openmp[version='11.0.1|>=11.0.1']
Package click conflicts for:
ipython -> black -> click[version='>=6.5|>=7.1.2|>=8.0.0']
rich-click -> click[version='>=7,<9']
click
Package tomlkit conflicts for:
meson-python -> tomli[version='>=1.0.0'] -> tomlkit
pytest -> tomli[version='>=1.0.0'] -> tomlkit
Package backports conflicts for:
pooch -> backports.tempfile -> backports
ipython -> backports.shutil_get_terminal_size -> backports
matplotlib -> backports.functools_lru_cache -> backports
Package sphinxcontrib-websupport conflicts for:
sphinx -> sphinxcontrib-websupport
numpydoc -> sphinx[version='>=1.8'] -> sphinxcontrib-websupport
pydata-sphinx-theme==0.9.0 -> sphinx -> sphinxcontrib-websupport
Package scandir conflicts for:
pytest -> pathlib2[version='>=2.2.0'] -> scandir
ipython -> pathlib2 -> scandir
Package meson conflicts for:
meson[version='>=0.64.0']
meson-python -> meson[version='>=0.60.0|>=0.62.0|>=0.63.3']
Package wheel conflicts for:
python -> pip -> wheel
meson-python -> wheel[version='>=0.36.0']
Package charset-normalizer conflicts for:
sphinx -> requests[version='>=2.25.0'] -> charset-normalizer[version='>=2,<3|>=2.0.0,<2.1|>=2.0.0,<2.0.1|>=2.0.0,<3|>=2.0.0,<2.1.0']
pooch -> requests[version='>=2.19.0'] -> charset-normalizer[version='>=2,<3|>=2.0.0,<2.1|>=2.0.0,<2.0.1|>=2.0.0,<3|>=2.0.0,<2.1.0']
Package attrs conflicts for:
pytest-timeout -> pytest[version='>=5.0.0'] -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
pytest-cov -> pytest[version='>=4.6'] -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
ipython -> black -> attrs[version='>=17.4.0|>=18.1.0']
pytest-xdist -> pytest[version='>=6.2.0'] -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
pytest -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
Package openblas conflicts for:
numpy -> libblas[version='>=3.8.0,<4.0a0'] -> openblas[version='0.3.5.*|0.3.6|>=0.3.6,<0.3.7.0a0',build=h828a276_2]
openblas
matplotlib -> numpy[version='>=1.14.6,<2.0a0'] -> openblas[version='0.2.20|0.2.20.*|>=0.2.20,<0.2.21.0a0|>=0.3.3,<0.3.4.0a0']
numpy -> openblas[version='0.2.20|0.2.20.*|>=0.2.20,<0.2.21.0a0|>=0.3.3,<0.3.4.0a0']
libblas[build=*openblas] -> openblas[version='0.3.5.*|0.3.6|>=0.3.6,<0.3.7.0a0',build=h828a276_2]
Package psutil conflicts for:
mypy -> psutil[version='>=4.0|>=5.4.0,<5.5.0']
pytest-xdist -> psutil[version='>=3.0']
Package cython conflicts for:
cython
cython-lint -> cython[version='>=0.29.32']
Package contextlib2 conflicts for:
pytest -> importlib-metadata[version='>=0.12'] -> contextlib2
click -> importlib-metadata -> contextlib2The following specifications were found to be incompatible with your system:
- feature:/win-64::__win==0=0
- feature:|@/win-64::__win==0=0
- click -> __unix
- click -> __win
- ipython -> __linux
- ipython -> __win
- rich-click -> click[version='>=7,<9'] -> __unix
- rich-click -> click[version='>=7,<9'] -> __win
- setuptools[version='<60.0'] -> wincertstore[version='>=0.2'] -> __win
Your installed version is: 0
```
### SciPy/NumPy/Python version information
Python 3.9.13 (main, Aug 25 2022, 23:51:50) [MSC v.1916 64 bit (AMD64)] :: Anaconda, Inc. on win32 Type "help", "copyright", "credits" or "license" for more information. >>> import scipy >>> scipy.show_config() lapack_mkl_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] lapack_opt_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] blas_mkl_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] blas_opt_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] >>> import sys >>> import numpy >>> print(scipy.__version__, numpy.__version__, sys.version_info) 1.9.1 1.21.5 sys.version_info(major=3, minor=9, micro=13, releaselevel='final', serial=0)
|
1.0
|
BUG: conda 23.1.0 environment creation fails on windows 11 - ### Describe your issue.
I am running into issues with conda setup as described in the [environment.yml file](https://github.com/scipy/scipy/blob/main/environment.yml). This is my first time trying to setup the scipy build environment, so I am not sure yet if this is something specific to me or a larger issue. My main clue at the moment is that the error message I am getting seems to imply several conflicts in the specified yml package version numbers - which I don't think I have control over locally.
Versions for reference:
scipy source commit ddc0d31b1
anaconda 23.1.0
Windows 11
The output under the scipy/numpy/Python version information block below was generated from the base Anaconda interpreter on my machine.
### Reproducing Code Example
```python
conda env create -f .\environment.yml
```
### Error message
```shell
Collecting package metadata (repodata.json): done
Solving environment: \
Found conflicts! Looking for incompatible packages.
This can take several minutes. Press CTRL-C to abort.
Examining pytest: 16%|███████████▋ | 6/38 [00:29<01:10, 2.22s/it]-Examining pydata-sphinx-theme==0.9.0: 45%|███████████████████████▋ | 17/38 [02:18<03:12, 9.18s/it]\Examining click: 63%|██████████████████████████████████████████████▋ | 24/38 [02:24<00:39, 2.81s/it]-Examining setuptools[version='<60.0']: 79%|█████████████████████████████████████████ | 30/38 [03:00<01:11, 8.94s/it]\Examining conflict for python setuptools ruff numpy click pybind11 ipython cython pytest asv pythran matplotlib: 45%|▍| 17/38 [|Examining conflict for sphinx pooch pydata-sphinx-theme pytest matplotlib: 61%|█████████▋ | 23/38 [02:31<00:54, 3.65s/it]/Examining conflict for pytest-cov sphinx-design sphinx python numpydoc pytest-xdist click ipython rich-click pytest pydata-sphin-
failed |
Solving environment: \
Found conflicts! Looking for incompatible packages.
This can take several minutes. Press CTRL-C to abort.
Examining pytest: 18%|█████████████▋ | 7/38 [00:37<03:55, 7.59s/it]|
Examining matplotlib: 24%|████████████████▌ | 9/38 [01:28<06:17, 13.03s/it]-Examining ipython: 37%|██████████████████████████▌ | 14/38 [01:53<02:14, 5.62s/it]\Examining click: 61%|████████████████████████████████████████████▊ | 23/38 [02:35<00:39, 2.62s/it]/Examining meson[version='>=0.64.0']: 68%|████████████████████████████████████▉ | 26/38 [02:37<00:22, 1.84s/it]\
Examining meson-python: 92%|█████████████████████████████████████████████████████████████▋ | 35/38 [03:26<00:10, 3.57s/it]/
Examining typing_extensions: 92%|█████████████████████████████████████████████████████████ | 35/38 [03:26<00:10, 3.57s/it]|Examining conflict for mpmath gmpy2: 5%|██▉ | 2/38 [00:14<04:22, 7.30s/it]\Examining conflict for mpmath openblas numpy pytest-xdist pytest asv matplotlib python cython pkg-config ninja click gmpy2 pythr\
Examining conflict for pytest pydevtool click: 24%|██████████▋ | 9/38 [01:33<05:12, 10.79s/it]\Examining conflict for numpy pythran libblas matplotlib: 42%|██████████████▎ | 16/38 [01:57<01:56, 5.30s/it]/
Examining conflict for sphinx pooch: 53%|████████████████████████████▍ | 20/38 [02:25<01:35, 5.29s/it]\Examining conflict for sphinx setuptools pooch pytest-xdist meson ipython cython pytest asv pytest-cov matplotlib: 68%|▋| 26/38/
Examining conflict for pytest-cov sphinx-design sphinx python setuptools cython-lint numpydoc pytest-xdist meson ipython cython /Examining conflict for sphinx setuptools pytest-xdist meson ipython cython pytest asv pytest-cov matplotlib: 89%|▉| 34/38 [02:5-
Examining conflict for pytest ipython: 95%|█████████████████████████████████████████████████▎ | 36/38 [03:07<00:04, 2.45s/it]/failed \ / /
UnsatisfiableError: The following specifications were found to be incompatible with each other:
Output in format: Requested package -> Available versions
Package appdirs conflicts for:
ipython -> black -> appdirs
pooch -> appdirs[version='>=1.3.0']
Package backports.functools_lru_cache conflicts for:
pytest -> wcwidth -> backports.functools_lru_cache
matplotlib -> backports.functools_lru_cache
Package colorama conflicts for:
meson-python -> colorama
numpydoc -> sphinx[version='>=1.8'] -> colorama[version='>=0.3.5|>=0.4.5']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> colorama[version='>=0.3.5|>=0.4.5']
ipython -> colorama
python -> pip -> colorama
click -> colorama
pytest -> colorama
pytest-cov -> pytest[version='>=4.6'] -> colorama
sphinx -> colorama[version='>=0.3.5|>=0.4.5']
pytest-timeout -> pytest[version='>=5.0.0'] -> colorama
rich-click -> click[version='>=7,<9'] -> colorama[version='>=0.4.0,<0.5.0']
sphinx-design -> sphinx[version='>=4,<6'] -> colorama[version='>=0.3.5|>=0.4.5']
pytest-xdist -> pytest[version='>=6.2.0'] -> colorama
Package vs2015_runtime conflicts for:
ipython -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
cython-lint -> cython[version='>=0.29.32'] -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0|>=14.16.27033|>=14.29.30139']
mypy -> psutil[version='>=4.0'] -> vs2015_runtime[version='>=14.15.26706|>=14.16.27012,<15.0a0|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
numpy -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
pooch -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
ruff -> vs2015_runtime[version='>=14.29.30139']
compilers -> c-compiler==1.2.0=h8ffe710_0 -> vs2015_runtime[version='>=14.16.27012']
doit[version='>=0.36.0'] -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
libblas[build=*openblas] -> libopenblas==0.3.21=pthreads_h02691f0_0 -> vs2015_runtime[version='>=14.16.27012|>=14.16.27033']
sphinx -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
numpy -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
rich-click -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pkg-config -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0']
typing_extensions -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
sphinx-design -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
setuptools[version='<60.0'] -> python[version='>=3.11,<3.12.0a0'] -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0|>=14.16.27033|>=14.29.30139']
pybind11 -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
meson-python -> ninja -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
asv[version='<0.5'] -> python[version='>=3.10,<3.11.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27012,<15.0a0|>=14.16.27033|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30139|>=14.29.30037|>=14.27.29016']
cython -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
ruff -> python[version='>=3.8,<3.9.0a0'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
pkg-config -> libglib[version='>=2.64.6,<3.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27033|>=14.29.30139|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
pybind11 -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
openblas -> vs2015_runtime[version='>=14.16.27012|>=14.16.27033|>=14.29.30139']
asv[version='<0.5'] -> vs2015_runtime[version='>=14.16.27012']
cython -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
numpydoc -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pytest-timeout -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pythran[version='>=0.11.0'] -> clang -> vs2015_runtime[version='>=14.16.27012|>=14.16.27012,<15.0a0|>=14.29.30139|>=14.16.27033']
ninja -> vc[version='>=14.1,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30139|>=14.29.30037']
gmpy2 -> python[version='>=3.11,<3.12.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27033|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
gmpy2 -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.29.30139']
pytest -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
python -> libsqlite[version='>=3.39.4,<4.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.29.30139|>=14.34.31931|>=14.32.31332|>=14.29.30037']
pydevtool -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
matplotlib -> vs2015_runtime[version='>=14.16.27012,<15.0a0']
openblas -> vc[version='>=14.2,<15'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.27.29016|>=14.28.29325|>=14.32.31332|>=14.34.31931|>=14.29.30037']
ninja -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
threadpoolctl -> python[version='>=3.6'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
pydata-sphinx-theme==0.9.0 -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
mpmath -> python[version='>=3'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
pytest-xdist -> python[version='>=3.7'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
mypy -> vs2015_runtime[version='>=14.16.27012|>=14.16.27033|>=14.29.30139']
pytest-cov -> coverage[version='>=5.2.1'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
click -> python[version='>=3.8'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033|>=14.29.30139']
python -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2'] -> vs2015_runtime[version='>=14.16.27012,<15.0a0|>=14.16.27012|>=14.16.27033']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> vs2015_runtime[version='>=14.0.25123,<15.0a0|>=14.0.25420|>=14.15.26706|>=14.16.27012|>=14.16.27033|>=14.29.30139|>=14.34.31931|>=14.32.31332|>=14.28.29325|>=14.29.30037|>=14.27.29016']
Package python conflicts for:
cython -> python_abi=3.9[build=*_cp39] -> python[version='3.10.*|3.11.*|3.8.*|3.9.*|3.7.*|>=3.7|>=3.8']
ninja -> python_abi=3.8[build=*_cp38] -> python[version='3.6.*|3.7.*|3.8.*']
pytest-xdist -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|>=3.7|3.4.*|>=2.7,<2.8.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
gmpy2 -> python_abi=3.11[build=*_cp311] -> python[version='3.10.*|3.11.*|3.9.*|3.8.*|3.7.*']
pooch -> appdirs[version='>=1.3.0'] -> python[version='2.7.*|3.5.*|3.6.*|3.4.*|>=3.7|>=2.7|>=3.9,<3.10.0a0|>=3.10,<3.11.0a0|>=3.7,<4.0|>=3.6,<4.0|>=3.11,<3.12.0a0']
pooch -> python[version='>=2.7,<2.8.0a0|>=3.5|>=3.6|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
pytest -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.8|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
click -> colorama -> python[version='3.10.*|3.8.*|>=3.7|3.11.*|3.9.*|3.7.*']
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2'] -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.7|>=3.8|>=3.11,<3.12.0a0|3.4.*']
pytest -> attrs[version='>=19.2.0'] -> python[version='2.7|>=3.4|3.10.*|3.8.*|3.9.*|>=2.7|>=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*|>=3.5|>=3.7|>=3.6|3.11.*|>=3|3.7.*|>=3.4|>=3.8.0a,<3.9.0a0']
python
pytest-cov -> coverage[version='>=5.2.1'] -> python[version='>=2.7|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.8|>=3.7']
numpydoc -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.7|3.4.*|>=3.7,<3.8.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.10,<3.11.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=2.7,<2.8.0a0']
rich-click -> click[version='>=7,<9'] -> python[version='>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.7.0|>=3.6.2']
ninja -> python[version='>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
sphinx-design -> sphinx[version='>=4,<6'] -> python[version='>=3.10,<3.11.0a0|>=3.7|>=3.7,<3.8.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.5']
mypy -> python[version='3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5|>=3.4|>=3.5,<3.6.0a0|>=3.6']
meson[version='>=0.64.0'] -> python[version='>=3.10,<3.11.0a0|>=3.5.2|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0']
numpy -> python[version='3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
cython-lint -> python[version='>=3.6|>=3.7']
pytest-cov -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|3.4.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.7,<3.8.0a0|>=3.5,<3.6.0a0']
doit[version='>=0.36.0'] -> python[version='>=3.8']
pythran[version='>=0.11.0'] -> python[version='>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0']
asv[version='<0.5'] -> python_abi=3.10[build=*_cp310] -> python[version='3.10.*|>=3.7|>=3.8|>=3.11,<3.12.0a0|3.9.*|3.7.*|3.8.*']
meson-python -> python[version='>=3.10,<3.11.0a0|>=3.7|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0']
ipython -> python[version='2.7.*|3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.8|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=2.7,<2.8.0a0|3.4.*']
sphinx -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.5|>=3.6|>=3.7|>=3.8|>=3.6,<3.7.0a0|>=3.7,<3.8.0a0|>=3.5,<3.6.0a0|3.4.*|>=3.9,<3.10.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0']
matplotlib -> python[version='3.4.*|3.5.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
cython -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
matplotlib -> tornado[version='>=5'] -> python[version='2.7.*|3.6.*|3.10.*|3.8.*|3.9.*|3.11.*|3.7.*|>=3.6|>=3.0,<=3.9|>=3|>=3.7|>=3.8|<3']
pydevtool -> python[version='>=3.8']
pydata-sphinx-theme==0.9.0 -> python[version='>=3.7']
sphinx -> alabaster[version='>=0.7,<0.8'] -> python[version='>=2.7|>=2|>=3.11,<3.12.0a0|>=3.4|>=3.7,<4.0|>=3.6,<4.0|>=3']
numpy -> python_abi=3.10[build=*_cp310] -> python[version='2.7.*|3.10.*|3.11.*|3.9.*|3.8.*|3.7.*']
threadpoolctl -> python[version='>=3.5|>=3.6']
mypy -> mypy_extensions[version='>=0.4.3'] -> python[version='2.7.*|>=2.7,<2.8.0a0|3.4.*|3.8.*|>=3.7|3.10.*|3.9.*|3.11.*|3.7.*|>=2.7|>=3']
typing_extensions -> python[version='>=2.7,<2.8.0a0|>=3.5|>=3.6|>=3.7|>=3.6,<3.7.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0|>=3.5,<3.6.0a0|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0']
ruff -> python_abi=3.8[build=*_cp38] -> python[version='3.10.*|3.8.*|3.11.*|3.9.*']
pytest-xdist -> execnet[version='>=1.1'] -> python[version='2.7|>=3.5|>=3.8|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=2.7']
setuptools[version='<60.0'] -> python_abi=3.11[build=*_cp311] -> python[version='3.10.*|3.11.*|3.8.*|3.7.*|3.9.*|>=3.7|>=2.7']
ruff -> python[version='>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0']
pydevtool -> doit -> python[version='3.5.*|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.6.*|>=2.7,<2.8.0a0']
pythran[version='>=0.11.0'] -> beniget=0.4 -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|3.4.*|>=3.6,<3.7.0a0|>=2.7,<2.8.0a0|>=3.5,<3.6.0a0|>=3.4|3.11.*|3.8.*|3.10.*|3.9.*|3.7.*']
pybind11 -> python_abi=3.10[build=*_cp310] -> python[version='3.10.*|3.8.*|3.11.*|3.9.*|3.7.*']
asv[version='<0.5'] -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
sphinx-design -> python[version='>=3.6']
gmpy2 -> python[version='3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
pytest-timeout -> python[version='2.7.*|3.4.*|3.5.*|3.6.*|>=3.6|>=3.8,<3.9.0a0|>=2.7,<2.8.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0']
doit[version='>=0.36.0'] -> cloudpickle -> python[version='2.7.*|3.5.*|3.6.*|>=3.5|>=3.6|3.4.*|>=3|>=2.7,<2.8.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.10,<3.11.0a0']
pybind11 -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
meson-python -> colorama -> python[version='!=3.0,!=3.1,!=3.2,!=3.3,!=3.4|2.7.*|3.5.*|3.6.*|>=3.6|3.4.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.5.2|2.7.*|>=3.6']
ipython -> backcall -> python[version='3.10.*|3.9.*|>=2.7|>=3|>=3.5|>=3.6|>=3.7|3.11.*|3.8.*|>=3.11,<3.12.0a0|3.7.*']
click -> python[version='2.7.*|3.5.*|3.6.*|>=3.10,<3.11.0a0|>=3.8|>=3.8,<3.9.0a0|>=3.11,<3.12.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|3.4.*|>=3|>=3.6|>=2.7,<2.8.0a0|>=3.5,<3.6.0a0']
typing_extensions -> python_abi=3.6[build=*_cp36m] -> python[version='2.7.*|3.6.*|3.7.*|3.8.*|>=3|3.5.*']
pytest-timeout -> pytest[version='>=5.0.0'] -> python[version='>=3.10,<3.11.0a0|>=3.8|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0']
cython-lint -> cython[version='>=0.29.32'] -> python[version='2.7.*|2.7.*|>=3.5|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|3.6.*|3.5.*|3.4.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|>=3.6.1']
rich-click -> python[version='>=3.6|>=3.7']
numpydoc -> jinja2[version='>=2.10'] -> python[version='>=3.6|>=3.8']
setuptools[version='<60.0'] -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.10,<3.11.0a0|>=3.11,<3.12.0a0|>=3.8,<3.9.0a0|>=3.9,<3.10.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0|>=3.5,<3.6.0a0|3.4.*']
pkg-config -> glib[version='>=2.58.3,<3.0a0'] -> python[version='*|>=2.7|>=3.8,<3.9.0a0|>=3.7,<3.8.0a0|>=3.6,<3.7.0a0']
pydata-sphinx-theme==0.9.0 -> beautifulsoup4 -> python[version='2.7.*|3.5.*|3.6.*|>=2.7,<2.8.0a0|>=3.6,<3.7.0a0|>=3.6|>=3.7,<3.8.0a0|>=3.8,<3.9.0a0|>=3.5,<3.6.0a0|>=3.10,<3.11.0a0|>=3.9,<3.10.0a0|>=3.11,<3.12.0a0|3.4.*|>=2.7|>=3.8|>=3.5']
Package six conflicts for:
ipython -> prompt-toolkit[version='>3.0.1,<3.1.0'] -> six[version='>=1.13.0|>=1.9.0']
asv[version='<0.5'] -> six
pydevtool -> doit -> six
pytest -> six[version='>=1.10.0']
sphinx -> six[version='>=1.5']
pydata-sphinx-theme==0.9.0 -> packaging -> six[version='>=1.5']
pooch -> packaging[version='>=20.0'] -> six
numpy -> mkl-service[version='>=2.3.0,<3.0a0'] -> six
pytest-xdist -> six
pytest-timeout -> pytest[version='>=3.6.0'] -> six[version='>=1.10.0']
numpydoc -> sphinx[version='>=1.8'] -> six[version='>=1.5']
pytest -> more-itertools[version='>=4.0.0'] -> six[version='>=1.0.0,<2.0.0|>=1.13.0']
pytest-xdist -> pytest[version='>=4.4.0'] -> six[version='>=1.10.0']
pytest-cov -> pytest[version='>=4.6'] -> six[version='>=1.10.0']
matplotlib -> cycler -> six[version='>=1.5']
Package typing_extensions conflicts for:
ipython -> black -> typing_extensions[version='>=3.10.0.0|>=3.10|>=3.10,!=3.10.0.1|>=3.7.4']
typing_extensions
click -> importlib-metadata -> typing_extensions[version='>=3.6.4']
rich-click -> rich[version='>=10'] -> typing_extensions[version='>=3.7.4,<4.0.0|>=3.7.4,<5.0.0|>=4.0.0,<5.0.0']
pytest -> importlib-metadata[version='>=0.12'] -> typing_extensions[version='>=3.6.4']
matplotlib -> kiwisolver -> typing_extensions
sphinx -> importlib-metadata[version='>=4.8'] -> typing_extensions[version='>=3.6.4']
mypy -> typing_extensions[version='>=3.10|>=3.7.4']
meson-python -> typing-extensions[version='>=3.7.4'] -> typing_extensions[version='3.10.0.0|3.10.0.0|3.10.0.1|3.10.0.2|4.0.0|4.0.1|4.1.1|4.2.0|4.2.0|4.3.0|4.4.0|3.7.4.3|3.7.4.2|3.7.4.1|3.7.4.1|3.7.4.1|3.7.4.1|3.7.4|4.4.0|4.4.0|4.4.0|4.4.0|4.3.0|4.3.0|4.3.0|4.3.0|4.1.1|3.10.0.2|3.7.4.3|3.7.4.1|3.7.4',build='py36_0|pyh06a4308_0|py38haa95532_0|py310haa95532_0|py37haa95532_0|py39haa95532_0|py38haa95532_0|py37haa95532_0|py37_0|py36_0|py36h9f0ad1d_1|py36h9f0ad1d_2|py_0|py_0|pyha770c72_0|pyha770c72_0|pyha770c72_0|pyha770c72_0|pyha770c72_0|pyha770c72_1|pyha770c72_0|pyha770c72_0|pyha770c72_0|py38h32f6830_3|py310haa95532_0|py39haa95532_0|pyh06a4308_0|pyh06a4308_0|pyh06a4308_0|py38_0']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4'] -> typing_extensions[version='>=3.6.4']
Package pypy3.7 conflicts for:
pytest -> pypy3.7[version='>=7.3.5|>=7.3.7']
ipython -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
pythran[version='>=0.11.0'] -> pypy3.7[version='>=7.3.7']
click -> pypy3.7[version='>=7.3.5|>=7.3.7']
pybind11 -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
asv[version='<0.5'] -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
sphinx -> docutils[version='>=0.18,<0.20'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
numpydoc -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pytest-timeout -> pytest[version='>=5.0.0'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
meson-python -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pydata-sphinx-theme==0.9.0 -> docutils!=0.17.0 -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
threadpoolctl -> python[version='>=3.6'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pybind11 -> pypy3.7[version='>=7.3.5|>=7.3.7']
pytest-xdist -> pytest[version='>=6.2.0'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
cython -> pypy3.7[version='>=7.3.5|>=7.3.7']
pythran[version='>=0.11.0'] -> numpy[version='>=1.20.3,<2.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*|>=7.3.5']
cython -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
mypy -> mypy_extensions[version='>=0.4.3'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
asv[version='<0.5'] -> pypy3.7[version='>=7.3.5|>=7.3.7']
pooch -> python[version='>=3.6'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.7']
cython-lint -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
python -> pypy3.7[version='7.3.5.*|7.3.7.*']
gmpy2 -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
python -> python_abi==3.7[build=*_pypy37_pp73] -> pypy3.7=7.3
numpy -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
ninja -> python -> pypy3.7[version='7.3.5.*|7.3.7.*']
pytest -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
setuptools[version='<60.0'] -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
numpy -> pypy3.7[version='>=7.3.5|>=7.3.7']
matplotlib -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4'] -> pypy3.7[version='>=7.3.5|>=7.3.7']
meson[version='>=0.64.0'] -> python[version='>=3.5.2'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.7|>=7.3.5']
mpmath -> python[version='>=3'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
ipython -> pypy3.7[version='>=7.3.5|>=7.3.7']
typing_extensions -> python[version='>=3.7'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
setuptools[version='<60.0'] -> pypy3.7[version='>=7.3.5|>=7.3.7']
matplotlib -> pypy3.7[version='>=7.3.5|>=7.3.7']
sphinx-design -> python[version='>=3.6'] -> pypy3.7[version='7.3.5.*|7.3.7.*']
pytest-cov -> coverage[version='>=5.2.1'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
click -> python[version='>=3.7,<3.8.0a0'] -> pypy3.7[version='7.3.*|7.3.5.*|7.3.7.*']
rich-click -> click[version='>=7,<9'] -> pypy3.7[version='7.3.5.*|7.3.7.*|>=7.3.5|>=7.3.7']
Package numpy conflicts for:
matplotlib -> numpy[version='1.10.*|1.11.*|>=1.14.6,<2.0a0']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> numpy[version='>=1.11.3,<2.0a0|>=1.15.4,<2.0a0|>=1.16.5,<2.0a0|>=1.16.6,<2.0a0|>=1.17|>=1.19.2,<2.0a0|>=1.19|>=1.20.3,<2.0a0|>=1.21.6,<2.0a0|>=1.23.5,<2.0a0|>=1.23.4,<2.0a0|>=1.21.5,<2.0a0|>=1.19.5,<2.0a0|>=1.18.5,<2.0a0|>=1.21.4,<2.0a0|>=1.21.2,<2.0a0|>=1.17.5,<2.0a0|>=1.19.4,<2.0a0']
numpy
pythran[version='>=0.11.0'] -> numpy[version='>=1.16.6,<2.0a0|>=1.18.5,<2.0a0|>=1.19.5,<2.0a0|>=1.20.3,<2.0a0|>=1.23.5,<2.0a0|>=1.21.6,<2.0a0|>=1.23.4,<2.0a0|>=1.21.5,<2.0a0|>=1.21.4,<2.0a0']
Package tomli conflicts for:
pytest -> tomli[version='>=1.0.0']
pytest-cov -> coverage[version='>=5.2.1'] -> tomli[version='>=1.0.0']
mypy -> tomli[version='>=1.1.0']
meson-python -> tomli[version='>=1.0.0']
pytest-timeout -> pytest[version='>=5.0.0'] -> tomli[version='>=1.0.0']
pytest-xdist -> pytest[version='>=6.2.0'] -> tomli[version='>=1.0.0']
ipython -> black -> tomli[version='>=0.2.6,<2.0.0|>=1.1.0']
Package setuptools conflicts for:
pytest -> setuptools[version='>=40.0']
meson[version='>=0.64.0'] -> setuptools
pytest-xdist -> pytest[version='>=4.4.0'] -> setuptools[version='>=40.0']
python -> pip -> setuptools
numpydoc -> jinja2[version='>=2.10'] -> setuptools
sphinx -> setuptools
matplotlib -> setuptools
ipython -> pygments[version='>=2.4.0'] -> setuptools
cython -> setuptools
pytest-cov -> setuptools
cython-lint -> pycodestyle -> setuptools
sphinx-design -> sphinx[version='>=4,<6'] -> setuptools
meson-python -> meson[version='>=0.63.3'] -> setuptools
setuptools[version='<60.0']
pytest-xdist -> setuptools
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> setuptools
pytest-cov -> pytest[version='>=4.6'] -> setuptools[version='>=40.0']
asv[version='<0.5'] -> setuptools
ipython -> setuptools[version='>=18.5']
pytest-timeout -> pytest[version='>=5.0.0'] -> setuptools[version='>=40.0']
Package pathlib2 conflicts for:
pytest-xdist -> pytest[version='>=4.4.0'] -> pathlib2[version='>=2.2.0']
pytest -> pathlib2[version='>=2.2.0']
click -> importlib-metadata -> pathlib2
ipython -> pathlib2
pytest-cov -> pytest[version='>=4.6'] -> pathlib2[version='>=2.2.0']
pytest-timeout -> pytest[version='>=5.0.0'] -> pathlib2[version='>=2.2.0']
Package vs2008_runtime conflicts for:
numpy -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
click -> python -> vs2008_runtime
matplotlib -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2'] -> vs2008_runtime
pytest-cov -> python -> vs2008_runtime
numpydoc -> python -> vs2008_runtime
pooch -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
asv[version='<0.5'] -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
python -> vc[version='>=9,<10.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
ninja -> vc=9 -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
meson-python -> ninja -> vs2008_runtime
ninja -> vs2008_runtime
pytest-timeout -> python -> vs2008_runtime
setuptools[version='<60.0'] -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
sphinx -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
pytest-xdist -> python -> vs2008_runtime
mpmath -> python -> vs2008_runtime
cython -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
pybind11 -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime[version='>=9.0.30729.1,<10.0a0']
python -> vs2008_runtime
pytest -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
ipython -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
typing_extensions -> python[version='>=2.7,<2.8.0a0'] -> vs2008_runtime
Package typed-ast conflicts for:
ipython -> black -> typed-ast[version='>=1.4.0|>=1.4.2']
mypy -> typed-ast[version='>=1.1.0,<1.2.0|>=1.2.0,<1.3.0|>=1.3.1,<1.4.0|>=1.4.0,<1.5.0|>=1.4.0,<2']
Package wincertstore conflicts for:
sphinx -> setuptools -> wincertstore[version='>=0.2']
matplotlib -> setuptools -> wincertstore[version='>=0.2']
meson[version='>=0.64.0'] -> setuptools -> wincertstore[version='>=0.2']
pytest-cov -> setuptools -> wincertstore[version='>=0.2']
pytest-xdist -> setuptools -> wincertstore[version='>=0.2']
asv[version='<0.5'] -> setuptools -> wincertstore[version='>=0.2']
ipython -> setuptools[version='>=18.5'] -> wincertstore[version='>=0.2']
setuptools[version='<60.0'] -> wincertstore[version='>=0.2']
pytest -> setuptools[version='>=40.0'] -> wincertstore[version='>=0.2']
cython -> setuptools -> wincertstore[version='>=0.2']
Package importlib-metadata conflicts for:
sphinx-design -> sphinx[version='>=4,<6'] -> importlib-metadata[version='>=4.4|>=4.8']
rich-click -> click[version='>=7,<9'] -> importlib-metadata
pytest-timeout -> pytest[version='>=5.0.0'] -> importlib-metadata[version='>=0.12']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> importlib-metadata[version='>=4.4|>=4.8']
pytest-cov -> pytest[version='>=4.6'] -> importlib-metadata[version='>=0.12']
pytest -> importlib_metadata[version='>=0.12'] -> importlib-metadata[version='>=1.1.3,<1.1.4.0a0|>=1.5.0,<1.5.1.0a0|>=1.5.2,<1.5.3.0a0|>=1.6.0,<1.6.1.0a0|>=1.6.1,<1.6.2.0a0|>=1.7.0,<1.7.1.0a0|>=2.0.0,<2.0.1.0a0|>=3.0.0,<3.0.1.0a0|>=3.1.0,<3.1.1.0a0|>=3.1.1,<3.1.2.0a0|>=3.10.0,<3.10.1.0a0|>=3.10.1,<3.10.2.0a0|>=4.0.1,<4.0.2.0a0|>=4.10.0,<4.10.1.0a0|>=4.10.1,<4.10.2.0a0|>=4.11.0,<4.11.1.0a0|>=4.11.1,<4.11.2.0a0|>=4.11.2,<4.11.3.0a0|>=4.11.3,<4.11.4.0a0|>=4.11.4,<4.11.5.0a0|>=5.0.0,<5.0.1.0a0|>=5.1.0,<5.1.1.0a0|>=5.2.0,<5.2.1.0a0|>=6.0.0,<6.0.1.0a0|>=4.9.0,<4.9.1.0a0|>=4.8.3,<4.8.4.0a0|>=4.8.2,<4.8.3.0a0|>=4.8.1,<4.8.2.0a0|>=4.8.0,<4.8.1.0a0|>=4.7.1,<4.7.2.0a0|>=4.7.0,<4.7.1.0a0|>=4.6.4,<4.6.5.0a0|>=4.6.3,<4.6.4.0a0|>=4.6.2,<4.6.3.0a0|>=4.6.1,<4.6.2.0a0|>=4.6.0,<4.6.1.0a0|>=4.5.0,<4.5.1.0a0|>=4.4.0,<4.4.1.0a0|>=4.3.1,<4.3.2.0a0|>=4.3.0,<4.3.1.0a0|>=4.2.0,<4.2.1.0a0|>=3.9.1,<3.9.2.0a0|>=3.9.0,<3.9.1.0a0|>=3.8.1,<3.8.2.0a0|>=3.8.0,<3.8.1.0a0|>=3.7.3,<3.7.4.0a0|>=3.7.2,<3.7.3.0a0|>=3.7.0,<3.7.1.0a0|>=3.6.0,<3.6.1.0a0|>=3.4.0,<3.4.1.0a0|>=3.3.0,<3.3.1.0a0']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4']
pydevtool -> doit -> importlib-metadata[version='>=4.4']
numpydoc -> sphinx[version='>=1.8'] -> importlib-metadata[version='>=4.4|>=4.8']
pytest -> importlib-metadata[version='>=0.12']
pytest-xdist -> pytest[version='>=6.2.0'] -> importlib-metadata[version='>=0.12']
sphinx -> importlib-metadata[version='>=4.4|>=4.8']
click -> importlib-metadata
Package vs2010_runtime conflicts for:
numpydoc -> python -> vs2010_runtime
matplotlib -> python=3.4 -> vs2010_runtime
pybind11 -> python=3.4 -> vs2010_runtime
ipython -> python=3.4 -> vs2010_runtime
pytest-timeout -> python -> vs2010_runtime
pytest-xdist -> python -> vs2010_runtime
asv[version='<0.5'] -> python=3.4 -> vs2010_runtime
sphinx -> python=3.4 -> vs2010_runtime
setuptools[version='<60.0'] -> python=3.4 -> vs2010_runtime
mpmath -> python[version='>=3'] -> vs2010_runtime
pytest -> python=3.4 -> vs2010_runtime
ninja -> python -> vs2010_runtime
mypy -> python[version='>=3.4'] -> vs2010_runtime
click -> python -> vs2010_runtime
pytest-cov -> python -> vs2010_runtime
python -> vs2010_runtime
cython -> python=3.4 -> vs2010_runtime
Package packaging conflicts for:
pytest-cov -> pytest[version='>=4.6'] -> packaging
meson-python -> pyproject-metadata[version='>=0.6.1'] -> packaging[version='>=19.0']
ipython -> black -> packaging[version='>=22.0']
sphinx -> packaging[version='>=21.0']
pooch -> packaging[version='>=20.0']
pytest -> packaging
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> packaging[version='>=20.0']
python -> pip -> packaging
sphinx-design -> sphinx[version='>=4,<6'] -> packaging[version='>=21.0']
pytest-xdist -> pytest[version='>=6.2.0'] -> packaging
pydata-sphinx-theme==0.9.0 -> packaging
pytest-timeout -> pytest[version='>=5.0.0'] -> packaging
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> packaging[version='>=21.0']
numpydoc -> sphinx[version='>=1.8'] -> packaging[version='>=21.0']
Package zipp conflicts for:
click -> importlib-metadata -> zipp[version='>=0.5']
sphinx -> importlib-metadata[version='>=4.8'] -> zipp[version='>=0.5']
doit[version='>=0.36.0'] -> importlib-metadata[version='>=4.4'] -> zipp[version='>=0.5']
pytest -> importlib-metadata[version='>=0.12'] -> zipp[version='>=0.5']
Package sqlite conflicts for:
gmpy2 -> python[version='>=3.11,<3.12.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
numpydoc -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
ruff -> python[version='>=3.8,<3.9.0a0'] -> sqlite[version='>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.39.2,<4.0a0|>=3.39.1,<4.0a0|>=3.38.2,<4.0a0']
doit[version='>=0.36.0'] -> python[version='>=3.8'] -> sqlite[version='>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0']
threadpoolctl -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
cython -> python[version='>=3.9,<3.10.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.39.1,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
cython-lint -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
matplotlib -> pypy3.8[version='>=7.3.9'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.2,<4.0a0|>=3.39.1,<4.0a0|>=3.39.2,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
setuptools[version='<60.0'] -> python[version='>=3.11,<3.12.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.2,<4.0a0|>=3.39.1,<4.0a0|>=3.39.2,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
python -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest-timeout -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
click -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.39.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
meson-python -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0']
pooch -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pybind11 -> python[version='>=3.10,<3.11.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.39.1,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pydata-sphinx-theme==0.9.0 -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0']
mypy -> python[version='>=3.8,<3.9.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pythran[version='>=0.11.0'] -> pypy3.8[version='>=7.3.9'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.2,<4.0a0|>=3.39.1,<4.0a0|>=3.39.2,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.38.5,<4.0a0|>=3.37.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0']
pydevtool -> python[version='>=3.8'] -> sqlite[version='>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0']
ninja -> python -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
ipython -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest-cov -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
meson[version='>=0.64.0'] -> python[version='>=3.5.2'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
asv[version='<0.5'] -> python[version='>=3.10,<3.11.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.30.0,<4.0a0|>=3.35.1,<4.0a0']
mpmath -> python[version='>=3'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest-xdist -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
sphinx -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
typing_extensions -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
pytest -> python[version='>=3.8'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.39.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
rich-click -> python[version='>=3.7'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
sphinx-design -> python[version='>=3.6'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.30.0,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
numpy -> python[version='>=3.10,<3.11.0a0'] -> sqlite[version='>=3.25.3,<4.0a0|>=3.26.0,<4.0a0|>=3.28.0,<4.0a0|>=3.30.0,<4.0a0|>=3.30.1,<4.0a0|>=3.32.3,<4.0a0|>=3.33.0,<4.0a0|>=3.34.0,<4.0a0|>=3.35.5,<4.0a0|>=3.36.0,<4.0a0|>=3.37.0,<4.0a0|>=3.37.1,<4.0a0|>=3.38.5,<4.0a0|>=3.40.1,<4.0a0|>=3.40.0,<4.0a0|>=3.39.3,<4.0a0|>=3.38.0,<4.0a0|>=3.39.2,<4.0a0|>=3.39.1,<4.0a0|>=3.38.2,<4.0a0|>=3.35.4,<4.0a0|>=3.31.1,<4.0a0|>=3.29.0,<4.0a0|>=3.27.2,<4.0a0|>=3.35.1,<4.0a0']
python -> pypy3.9=7.3.11 -> sqlite[version='>=3.39.1,<4.0a0']
Package certifi conflicts for:
pytest-cov -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
sphinx -> requests[version='>=2.25.0'] -> certifi[version='>=2016.09|>=2016.9.26|>=2017.4.17']
asv[version='<0.5'] -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
meson[version='>=0.64.0'] -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
setuptools[version='<60.0'] -> certifi[version='>=2016.09|>=2016.9.26']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> certifi[version='>=2016.09|>=2016.9.26|>=2020.06.20']
cython -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
pytest -> setuptools[version='>=40.0'] -> certifi[version='>=2016.09|>=2016.9.26']
pytest-xdist -> setuptools -> certifi[version='>=2016.09|>=2016.9.26']
ipython -> setuptools[version='>=18.5'] -> certifi[version='>=2016.09|>=2016.9.26']
pooch -> requests[version='>=2.19.0'] -> certifi[version='>=2017.4.17']
Package configparser conflicts for:
pytest -> importlib-metadata[version='>=0.12'] -> configparser[version='>=3.5']
pydevtool -> doit -> configparser
click -> importlib-metadata -> configparser[version='>=3.5']
Package libflang conflicts for:
compilers -> fortran-compiler==1.2.0=h9655429_0 -> libflang[version='<6.0.0.a0|>=5.0.0|>=5.0.0,<6.0.0.a0']
libblas[build=*openblas] -> libopenblas==0.3.21=pthreads_h02691f0_0 -> libflang[version='<6.0.0.a0|>=5.0.0|>=5.0.0,<6.0.0.a0']
numpy -> libflang[version='<6.0.0.a0|>=5.0.0']
matplotlib -> numpy[version='>=1.14.6,<2.0a0'] -> libflang[version='<6.0.0.a0|>=5.0.0']
openblas -> libflang[version='<6.0.0.a0|>=5.0.0|>=5.0.0,<6.0.0.a0']
numpy -> openblas[version='>=0.2.20,<0.2.21.0a0'] -> libflang
Package sphinx conflicts for:
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2']
sphinx
numpydoc -> sphinx[version='>=1.8|>=4.2|>=3.0']
sphinx-design -> sphinx[version='>=3,<5|>=4,<6']
Package libopenblas conflicts for:
numpy -> libblas[version='>=3.9.0,<4.0a0'] -> libopenblas[version='0.3.10|0.3.12|0.3.12|0.3.15|0.3.17|0.3.18|0.3.20|0.3.21|0.3.9|0.3.8|0.3.7|0.3.7|0.3.6',build='hcfe7411_0|pthreads_h1662909_0|pthreads_h02691f0_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_h1662909_1|hc8a65f3_0|hcfe7411_0|h29e5d5d_1|h29e5d5d_0|h29e5d5d_6']
libblas[build=*openblas] -> openblas[version='>=0.3.6,<0.3.7.0a0'] -> libopenblas==0.3.6[build='h29e5d5d_3|h29e5d5d_4|h29e5d5d_5']
libblas[build=*openblas] -> libopenblas[version='0.3.10|0.3.12|0.3.12|0.3.15|0.3.17|0.3.18|0.3.20|0.3.21|0.3.9|0.3.8|0.3.7|0.3.7|0.3.6',build='hcfe7411_0|pthreads_h1662909_0|pthreads_h02691f0_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_h1662909_1|hc8a65f3_0|hcfe7411_0|h29e5d5d_1|h29e5d5d_0|h29e5d5d_6']
openblas -> libopenblas[version='0.3.10|0.3.10|0.3.10|0.3.10|0.3.10|0.3.10|0.3.12|0.3.12|0.3.13|0.3.15|0.3.15|0.3.16|0.3.17|0.3.17|0.3.18|0.3.20|0.3.20|0.3.21|0.3.9|0.3.8|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.7|0.3.6|0.3.6|0.3.6|0.3.6',build='h29e5d5d_3|h29e5d5d_4|h29e5d5d_5|hcfe7411_4|hcfe7411_7|hcfe7411_0|pthreads_h1662909_4|pthreads_hc469a61_1|pthreads_h02691f0_2|pthreads_hc140b1d_3|pthreads_h02691f0_1|pthreads_h02691f0_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_1|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_hc469a61_0|pthreads_h1662909_1|pthreads_h1662909_0|pthreads_h1662909_5|pthreads_h1662909_3|pthreads_h1662909_2|pthreads_h1662909_1|hc8a65f3_0|hcfe7411_0|hcfe7411_6|hcfe7411_5|h29e5d5d_3|h29e5d5d_2|h29e5d5d_1|h29e5d5d_0|h29e5d5d_6']
Package ca-certificates conflicts for:
matplotlib -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
pytest-timeout -> python -> ca-certificates
asv[version='<0.5'] -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
python -> ca-certificates
sphinx -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
setuptools[version='<60.0'] -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
numpy -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
mpmath -> python -> ca-certificates
pytest -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
ninja -> python -> ca-certificates
ipython -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
cython -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
pybind11 -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
numpydoc -> python -> ca-certificates
pooch -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
typing_extensions -> python[version='>=2.7,<2.8.0a0'] -> ca-certificates
pytest-xdist -> python -> ca-certificates
click -> python -> ca-certificates
pytest-cov -> python -> ca-certificates
Package pytest conflicts for:
pytest
pytest-xdist -> pytest[version='>=2.4.2|>=2.7.0|>=3.0.0|>=4.4.0|>=6.0.0|>=6.2.0']
pytest-xdist -> pytest-forked -> pytest[version='>=2.6.0|>=3.1.0|>=3.10']
pytest-cov -> pytest[version='>=3.6|>=4.6']
pytest-timeout -> pytest[version='>=2.8.0|>=3.6.0|>=5.0.0']
Package numpydoc conflicts for:
ipython -> jedi[version='>=0.10,<0.18'] -> numpydoc
numpydoc
Package expat conflicts for:
pythran[version='>=0.11.0'] -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
pytest -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
matplotlib -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
cython -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
numpy -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0|>=2.4.7,<3.0a0']
ipython -> pypy3.7[version='>=7.3.7'] -> expat[version='>=2.4.1,<3.0a0']
click -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
python -> pypy3.9=7.3.11 -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
setuptools[version='<60.0'] -> pypy3.8[version='>=7.3.8'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
ruff -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
asv[version='<0.5'] -> pypy3.7[version='>=7.3.7'] -> expat[version='>=2.4.1,<3.0a0']
pybind11 -> pypy3.8[version='>=7.3.9'] -> expat[version='>=2.4.1,<3.0a0|>=2.4.7,<3.0a0|>=2.4.8,<3.0a0|>=2.4.9,<3.0a0|>=2.5.0,<3.0a0']
Package typing conflicts for:
sphinx -> typing
pydata-sphinx-theme==0.9.0 -> sphinx -> typing
mypy -> typing[version='>=3.5.3']
typing_extensions -> typing[version='>=3.6.2|>=3.7.4']
mypy -> typing_extensions[version='>=3.7.4'] -> typing[version='>=3.7.4']
numpydoc -> sphinx[version='>=1.8'] -> typing
Package typing-extensions conflicts for:
meson-python -> typing-extensions[version='>=3.7.4']
matplotlib -> kiwisolver -> typing-extensions
Package ucrt conflicts for:
pytest-cov -> coverage[version='>=5.2.1'] -> ucrt[version='>=10.0.20348.0']
openblas -> vs2015_runtime[version='>=14.29.30139'] -> ucrt[version='>=10.0.20348.0']
gmpy2 -> ucrt[version='>=10.0.20348.0']
cython-lint -> cython[version='>=0.29.32'] -> ucrt[version='>=10.0.20348.0']
click -> pypy3.8[version='>=7.3.9'] -> ucrt[version='>=10.0.20348.0']
setuptools[version='<60.0'] -> pypy3.8[version='>=7.3.8'] -> ucrt[version='>=10.0.20348.0']
pytest -> pypy3.8[version='>=7.3.9'] -> ucrt[version='>=10.0.20348.0']
ruff -> ucrt[version='>=10.0.20348.0']
pytest-xdist -> psutil[version='>=3.0'] -> ucrt[version='>=10.0.20348.0']
pkg-config -> libglib[version='>=2.64.6,<3.0a0'] -> ucrt[version='>=10.0.20348.0']
mypy -> ucrt[version='>=10.0.20348.0']
mpmath -> gmpy2 -> ucrt[version='>=10.0.20348.0']
pythran[version='>=0.11.0'] -> numpy[version='>=1.20.3,<2.0a0'] -> ucrt[version='>=10.0.20348.0']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> ucrt[version='>=10.0.20348.0']
pybind11 -> ucrt[version='>=10.0.20348.0']
ninja -> vs2015_runtime[version='>=14.16.27033'] -> ucrt[version='>=10.0.20348.0']
cython -> ucrt[version='>=10.0.20348.0']
python -> libsqlite[version='>=3.39.4,<4.0a0'] -> ucrt[version='>=10.0.20348.0']
numpy -> ucrt[version='>=10.0.20348.0']
asv[version='<0.5'] -> vs2015_runtime[version='>=14.16.27012'] -> ucrt[version='>=10.0.20348.0']
Package dataclasses conflicts for:
ipython -> black -> dataclasses[version='>=0.6']
rich-click -> rich[version='>=10'] -> dataclasses[version='>=0.7,<0.9']
Package flit-core conflicts for:
typing_extensions -> flit-core[version='>=3.6,<4']
mypy -> typing_extensions[version='>=3.10'] -> flit-core[version='>=3.6,<4']
Package doit conflicts for:
doit[version='>=0.36.0']
pydevtool -> doit
Package pytz conflicts for:
sphinx -> babel[version='>=2.9'] -> pytz[version='>=2015.7']
matplotlib -> pytz
Package requests conflicts for:
python -> pip -> requests
sphinx -> requests[version='>2.0.0|>=2.0.0|>=2.25.0|>=2.5.0']
sphinx-design -> sphinx[version='>=4,<6'] -> requests[version='>=2.5.0']
pooch -> requests[version='>=2.19.0']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> requests[version='>2.0.0|>=2.0.0|>=2.25.0|>=2.5.0']
numpydoc -> sphinx[version='>=1.8'] -> requests[version='>2.0.0|>=2.0.0|>=2.25.0|>=2.5.0']
Package libblas conflicts for:
numpy -> libblas[version='>=3.8.0,<4.0a0|>=3.9.0,<4.0a0']
pythran[version='>=0.11.0'] -> numpy[version='>=1.20.3,<2.0a0'] -> libblas[version='>=3.8.0,<4.0a0|>=3.9.0,<4.0a0']
numpy -> libcblas[version='>=3.9.0,<4.0a0'] -> libblas[version='3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0.*|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.8.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0|3.9.0.*',build='5_openblas|5_blis|6_blis|6_openblas|7_blis|8_blis|8_openblas|8_mkl|9_blis|10_openblas|11_mkl|11_openblas|12_blis|13_mkl|14_openblas|15_blis|16_openblas|16_mkl|17_openblas|18_mkl|0_blis|0_openblas|1_openblas|2_openblas|4_blis|5_blis|5_openblas|5_mkl|6_blis|8_openblas|8_mkl|9_blis|9_openblas|10_openblas|11_win64_blis|13_win64_openblas|13_win64_mkl|14_win64_mkl|16_win64_openblas|16_win64_blis|16_win64_mkl|15_win64_blis|15_win64_mkl|15_win64_openblas|14_win64_openblas|14_win64_blis|13_win64_blis|12_win64_mkl|12_win64_openblas|12_win64_blis|11_win64_mkl|11_win64_openblas|10_mkl|10_blis|9_mkl|8_blis|7_mkl|7_openblas|7_blis|6_mkl|6_openblas|4_mkl|4_openblas|3_openblas|3_blis|2_blis|1_blis|21_mkl|20_mkl|19_mkl|16_blis|15_openblas|15_mkl|14_mkl|14_blis|13_openblas|13_blis|12_mkl|12_openblas|11_blis|10_mkl|10_blis|9_mkl|9_openblas|7_openblas|7_mkl|6_mkl|5_mkl|4_mkl|4_openblas|4_blis']
matplotlib -> numpy=1.11 -> libblas[version='>=3.8.0,<4.0a0|>=3.9.0,<4.0a0']
libblas[build=*openblas]
Package pygments conflicts for:
sphinx -> pygments[version='>2.0|>=2.0|>=2.12|>=2.13']
rich-click -> rich[version='>=10'] -> pygments[version='>=2.14.0,<3.0.0|>=2.6.0,<3.0.0']
pydata-sphinx-theme==0.9.0 -> sphinx[version='>=4.0.2'] -> pygments[version='>2.0|>=2.0|>=2.12|>=2.13']
ipython -> pygments[version='>=2.4.0']
numpydoc -> sphinx[version='>=1.8'] -> pygments[version='>2.0|>=2.0|>=2.12|>=2.13']
sphinx-design -> sphinx[version='>=4,<6'] -> pygments[version='>=2.0|>=2.12']
Package pyparsing conflicts for:
pooch -> packaging[version='>=20.0'] -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
matplotlib -> pyparsing
pydata-sphinx-theme==0.9.0 -> packaging -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
sphinx -> packaging[version='>=21.0'] -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
pytest -> packaging -> pyparsing[version='<3,>=2.0.2|>=2.0.2,!=3.0.5|>=2.0.2,<3|>=2.0.2']
matplotlib -> matplotlib-base[version='>=3.6.3,<3.6.4.0a0'] -> pyparsing[version='>=2.0.3,!=2.0.4,!=2.1.2,!=2.1.6|>=2.2.1|>=2.3.1']
Package ninja conflicts for:
meson[version='>=0.64.0'] -> ninja[version='>=1.8.2']
meson-python -> ninja
meson-python -> meson[version='>=0.63.3'] -> ninja[version='>=1.8.2']
ninja
Package decorator conflicts for:
ipython -> decorator
pythran[version='>=0.11.0'] -> decorator
Package llvm-openmp conflicts for:
openblas -> libflang[version='>=5.0.0'] -> llvm-openmp[version='11.0.1|>=11.0.1']
numpy -> libflang[version='>=5.0.0'] -> llvm-openmp[version='11.0.1|>=11.0.1']
Package click conflicts for:
ipython -> black -> click[version='>=6.5|>=7.1.2|>=8.0.0']
rich-click -> click[version='>=7,<9']
click
Package tomlkit conflicts for:
meson-python -> tomli[version='>=1.0.0'] -> tomlkit
pytest -> tomli[version='>=1.0.0'] -> tomlkit
Package backports conflicts for:
pooch -> backports.tempfile -> backports
ipython -> backports.shutil_get_terminal_size -> backports
matplotlib -> backports.functools_lru_cache -> backports
Package sphinxcontrib-websupport conflicts for:
sphinx -> sphinxcontrib-websupport
numpydoc -> sphinx[version='>=1.8'] -> sphinxcontrib-websupport
pydata-sphinx-theme==0.9.0 -> sphinx -> sphinxcontrib-websupport
Package scandir conflicts for:
pytest -> pathlib2[version='>=2.2.0'] -> scandir
ipython -> pathlib2 -> scandir
Package meson conflicts for:
meson[version='>=0.64.0']
meson-python -> meson[version='>=0.60.0|>=0.62.0|>=0.63.3']
Package wheel conflicts for:
python -> pip -> wheel
meson-python -> wheel[version='>=0.36.0']
Package charset-normalizer conflicts for:
sphinx -> requests[version='>=2.25.0'] -> charset-normalizer[version='>=2,<3|>=2.0.0,<2.1|>=2.0.0,<2.0.1|>=2.0.0,<3|>=2.0.0,<2.1.0']
pooch -> requests[version='>=2.19.0'] -> charset-normalizer[version='>=2,<3|>=2.0.0,<2.1|>=2.0.0,<2.0.1|>=2.0.0,<3|>=2.0.0,<2.1.0']
Package attrs conflicts for:
pytest-timeout -> pytest[version='>=5.0.0'] -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
pytest-cov -> pytest[version='>=4.6'] -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
ipython -> black -> attrs[version='>=17.4.0|>=18.1.0']
pytest-xdist -> pytest[version='>=6.2.0'] -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
pytest -> attrs[version='>=17.2.0|>=17.4.0,<=19.1.0|>=17.4.0|>=19.2.0']
Package openblas conflicts for:
numpy -> libblas[version='>=3.8.0,<4.0a0'] -> openblas[version='0.3.5.*|0.3.6|>=0.3.6,<0.3.7.0a0',build=h828a276_2]
openblas
matplotlib -> numpy[version='>=1.14.6,<2.0a0'] -> openblas[version='0.2.20|0.2.20.*|>=0.2.20,<0.2.21.0a0|>=0.3.3,<0.3.4.0a0']
numpy -> openblas[version='0.2.20|0.2.20.*|>=0.2.20,<0.2.21.0a0|>=0.3.3,<0.3.4.0a0']
libblas[build=*openblas] -> openblas[version='0.3.5.*|0.3.6|>=0.3.6,<0.3.7.0a0',build=h828a276_2]
Package psutil conflicts for:
mypy -> psutil[version='>=4.0|>=5.4.0,<5.5.0']
pytest-xdist -> psutil[version='>=3.0']
Package cython conflicts for:
cython
cython-lint -> cython[version='>=0.29.32']
Package contextlib2 conflicts for:
pytest -> importlib-metadata[version='>=0.12'] -> contextlib2
click -> importlib-metadata -> contextlib2The following specifications were found to be incompatible with your system:
- feature:/win-64::__win==0=0
- feature:|@/win-64::__win==0=0
- click -> __unix
- click -> __win
- ipython -> __linux
- ipython -> __win
- rich-click -> click[version='>=7,<9'] -> __unix
- rich-click -> click[version='>=7,<9'] -> __win
- setuptools[version='<60.0'] -> wincertstore[version='>=0.2'] -> __win
Your installed version is: 0
```
### SciPy/NumPy/Python version information
Python 3.9.13 (main, Aug 25 2022, 23:51:50) [MSC v.1916 64 bit (AMD64)] :: Anaconda, Inc. on win32 Type "help", "copyright", "credits" or "license" for more information. >>> import scipy >>> scipy.show_config() lapack_mkl_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] lapack_opt_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] blas_mkl_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] blas_opt_info: libraries = ['mkl_rt'] library_dirs = ['C:/Users/Stephen/anaconda3\\Library\\lib'] define_macros = [('SCIPY_MKL_H', None), ('HAVE_CBLAS', None)] include_dirs = ['C:/Users/Stephen/anaconda3\\Library\\include'] >>> import sys >>> import numpy >>> print(scipy.__version__, numpy.__version__, sys.version_info) 1.9.1 1.21.5 sys.version_info(major=3, minor=9, micro=13, releaselevel='final', serial=0)
|
defect
|
bug conda environment creation fails on windows describe your issue i am running into issues with conda setup as described in the this is my first time trying to setup the scipy build environment so i am not sure yet if this is something specific to me or a larger issue my main clue at the moment is that the error message i am getting seems to imply several conflicts in the specified yml package version numbers which i don t think i have control over locally versions for reference scipy source commit anaconda windows the output under the scipy numpy python version information block below was generated from the base anaconda interpreter on my machine reproducing code example python conda env create f environment yml error message shell collecting package metadata repodata json done solving environment found conflicts looking for incompatible packages this can take several minutes press ctrl c to abort examining pytest ███████████▋ examining pydata sphinx theme ███████████████████████▋ examining click ██████████████████████████████████████████████▋ examining setuptools █████████████████████████████████████████ examining conflict for python setuptools ruff numpy click ipython cython pytest asv pythran matplotlib ▍ examining conflict for pytest cov sphinx design sphinx python numpydoc pytest xdist click ipython rich click pytest pydata sphin failed solving environment found conflicts looking for incompatible packages this can take several minutes press ctrl c to abort examining pytest █████████████▋ examining matplotlib ████████████████▌ examining ipython ██████████████████████████▌ examining click ████████████████████████████████████████████▊ examining meson ████████████████████████████████████▉ examining meson python █████████████████████████████████████████████████████████████▋ examining typing extensions █████████████████████████████████████████████████████████ examining conflict for mpmath ██▉ examining conflict for mpmath openblas numpy pytest xdist pytest asv matplotlib python cython pkg config ninja click pythr examining conflict for pytest pydevtool click ██████████▋ examining conflict for numpy pythran libblas matplotlib ██████████████▎ examining conflict for sphinx pooch ████████████████████████████▍ examining conflict for sphinx setuptools pooch pytest xdist meson ipython cython pytest asv pytest cov matplotlib ▋ examining conflict for pytest cov sphinx design sphinx python setuptools cython lint numpydoc pytest xdist meson ipython cython examining conflict for sphinx setuptools pytest xdist meson ipython cython pytest asv pytest cov matplotlib ▉ examining conflict for pytest ipython █████████████████████████████████████████████████▎ failed unsatisfiableerror the following specifications were found to be incompatible with each other output in format requested package available versions package appdirs conflicts for ipython black appdirs pooch appdirs package backports functools lru cache conflicts for pytest wcwidth backports functools lru cache matplotlib backports functools lru cache package colorama conflicts for meson python colorama numpydoc sphinx colorama pydata sphinx theme sphinx colorama ipython colorama python pip colorama click colorama pytest colorama pytest cov pytest colorama sphinx colorama pytest timeout pytest colorama rich click click colorama sphinx design sphinx colorama pytest xdist pytest colorama package runtime conflicts for ipython python runtime cython lint cython runtime mypy psutil runtime numpy runtime pooch python runtime ruff runtime compilers c compiler runtime doit python runtime libblas libopenblas pthreads runtime sphinx python runtime numpy vc runtime rich click python runtime pkg config runtime typing extensions python runtime sphinx design python runtime setuptools python runtime vc runtime meson python ninja runtime asv python runtime cython vc runtime ruff python runtime pkg config libglib runtime runtime openblas runtime asv runtime cython runtime numpydoc python runtime pytest timeout python runtime pythran clang runtime ninja vc runtime python runtime runtime pytest python runtime python libsqlite runtime pydevtool python runtime matplotlib runtime openblas vc runtime ninja runtime threadpoolctl python runtime pydata sphinx theme python runtime mpmath python runtime pytest xdist python runtime mypy runtime pytest cov coverage runtime click python runtime python runtime meson ninja runtime matplotlib matplotlib base runtime package python conflicts for cython python abi python ninja python abi python pytest xdist python python abi python pooch appdirs python pooch python pytest python click colorama python meson ninja python pytest attrs python python pytest cov coverage python numpydoc python rich click click python ninja python sphinx design sphinx python mypy python meson python numpy python cython lint python pytest cov python doit python pythran python asv python abi python meson python python ipython python sphinx python matplotlib python cython python matplotlib tornado python pydevtool python pydata sphinx theme python sphinx alabaster python numpy python abi python threadpoolctl python mypy mypy extensions python typing extensions python ruff python abi python pytest xdist execnet python setuptools python abi python ruff python pydevtool doit python pythran beniget python python abi python asv python sphinx design python python pytest timeout python doit cloudpickle python python meson python colorama python ipython backcall python click python typing extensions python abi python pytest timeout pytest python cython lint cython python rich click python numpydoc python setuptools python pkg config glib python pydata sphinx theme python package six conflicts for ipython prompt toolkit six asv six pydevtool doit six pytest six sphinx six pydata sphinx theme packaging six pooch packaging six numpy mkl service six pytest xdist six pytest timeout pytest six numpydoc sphinx six pytest more itertools six pytest xdist pytest six pytest cov pytest six matplotlib cycler six package typing extensions conflicts for ipython black typing extensions typing extensions click importlib metadata typing extensions rich click rich typing extensions pytest importlib metadata typing extensions matplotlib kiwisolver typing extensions sphinx importlib metadata typing extensions mypy typing extensions meson python typing extensions typing extensions doit importlib metadata typing extensions package conflicts for pytest ipython python pythran click python asv python sphinx docutils numpydoc python pytest timeout pytest meson python python pydata sphinx theme docutils threadpoolctl python pytest xdist pytest cython pythran numpy cython python mypy mypy extensions asv pooch python cython lint python python python python python abi numpy python ninja python pytest python setuptools python numpy matplotlib python doit importlib metadata meson python mpmath python ipython typing extensions python setuptools matplotlib sphinx design python pytest cov coverage click python rich click click package numpy conflicts for matplotlib numpy matplotlib matplotlib base numpy numpy pythran numpy package tomli conflicts for pytest tomli pytest cov coverage tomli mypy tomli meson python tomli pytest timeout pytest tomli pytest xdist pytest tomli ipython black tomli package setuptools conflicts for pytest setuptools meson setuptools pytest xdist pytest setuptools python pip setuptools numpydoc setuptools sphinx setuptools matplotlib setuptools ipython pygments setuptools cython setuptools pytest cov setuptools cython lint pycodestyle setuptools sphinx design sphinx setuptools meson python meson setuptools setuptools pytest xdist setuptools pydata sphinx theme sphinx setuptools pytest cov pytest setuptools asv setuptools ipython setuptools pytest timeout pytest setuptools package conflicts for pytest xdist pytest pytest click importlib metadata ipython pytest cov pytest pytest timeout pytest package runtime conflicts for numpy python runtime click python runtime matplotlib python runtime meson ninja runtime pytest cov python runtime numpydoc python runtime pooch python runtime asv python runtime python vc runtime ninja vc runtime meson python ninja runtime ninja runtime pytest timeout python runtime setuptools python runtime sphinx python runtime pytest xdist python runtime mpmath python runtime cython python runtime python runtime python runtime pytest python runtime ipython python runtime typing extensions python runtime package typed ast conflicts for ipython black typed ast mypy typed ast package wincertstore conflicts for sphinx setuptools wincertstore matplotlib setuptools wincertstore meson setuptools wincertstore pytest cov setuptools wincertstore pytest xdist setuptools wincertstore asv setuptools wincertstore ipython setuptools wincertstore setuptools wincertstore pytest setuptools wincertstore cython setuptools wincertstore package importlib metadata conflicts for sphinx design sphinx importlib metadata rich click click importlib metadata pytest timeout pytest importlib metadata pydata sphinx theme sphinx importlib metadata pytest cov pytest importlib metadata pytest importlib metadata importlib metadata doit importlib metadata pydevtool doit importlib metadata numpydoc sphinx importlib metadata pytest importlib metadata pytest xdist pytest importlib metadata sphinx importlib metadata click importlib metadata package runtime conflicts for numpydoc python runtime matplotlib python runtime python runtime ipython python runtime pytest timeout python runtime pytest xdist python runtime asv python runtime sphinx python runtime setuptools python runtime mpmath python runtime pytest python runtime ninja python runtime mypy python runtime click python runtime pytest cov python runtime python runtime cython python runtime package packaging conflicts for pytest cov pytest packaging meson python pyproject metadata packaging ipython black packaging sphinx packaging pooch packaging pytest packaging matplotlib matplotlib base packaging python pip packaging sphinx design sphinx packaging pytest xdist pytest packaging pydata sphinx theme packaging pytest timeout pytest packaging pydata sphinx theme sphinx packaging numpydoc sphinx packaging package zipp conflicts for click importlib metadata zipp sphinx importlib metadata zipp doit importlib metadata zipp pytest importlib metadata zipp package sqlite conflicts for python sqlite numpydoc python sqlite ruff python sqlite doit python sqlite threadpoolctl python sqlite cython python sqlite cython lint python sqlite matplotlib sqlite setuptools python sqlite python sqlite pytest timeout python sqlite click python sqlite meson python python sqlite pooch python sqlite python sqlite pydata sphinx theme python sqlite mypy python sqlite pythran sqlite pydevtool python sqlite ninja python sqlite ipython python sqlite pytest cov python sqlite meson python sqlite asv python sqlite mpmath python sqlite pytest xdist python sqlite sphinx python sqlite typing extensions python sqlite pytest python sqlite rich click python sqlite sphinx design python sqlite numpy python sqlite python sqlite package certifi conflicts for pytest cov setuptools certifi sphinx requests certifi asv setuptools certifi meson setuptools certifi setuptools certifi matplotlib matplotlib base certifi cython setuptools certifi pytest setuptools certifi pytest xdist setuptools certifi ipython setuptools certifi pooch requests certifi package configparser conflicts for pytest importlib metadata configparser pydevtool doit configparser click importlib metadata configparser package libflang conflicts for compilers fortran compiler libflang libblas libopenblas pthreads libflang numpy libflang matplotlib numpy libflang openblas libflang numpy openblas libflang package sphinx conflicts for pydata sphinx theme sphinx sphinx numpydoc sphinx sphinx design sphinx package libopenblas conflicts for numpy libblas libopenblas libblas openblas libopenblas libblas libopenblas openblas libopenblas package ca certificates conflicts for matplotlib python ca certificates pytest timeout python ca certificates asv python ca certificates python ca certificates sphinx python ca certificates setuptools python ca certificates numpy python ca certificates mpmath python ca certificates pytest python ca certificates ninja python ca certificates ipython python ca certificates cython python ca certificates python ca certificates numpydoc python ca certificates pooch python ca certificates typing extensions python ca certificates pytest xdist python ca certificates click python ca certificates pytest cov python ca certificates package pytest conflicts for pytest pytest xdist pytest pytest xdist pytest forked pytest pytest cov pytest pytest timeout pytest package numpydoc conflicts for ipython jedi numpydoc numpydoc package expat conflicts for pythran expat pytest expat matplotlib expat cython expat numpy expat ipython expat click expat python expat setuptools expat ruff expat asv expat expat package typing conflicts for sphinx typing pydata sphinx theme sphinx typing mypy typing typing extensions typing mypy typing extensions typing numpydoc sphinx typing package typing extensions conflicts for meson python typing extensions matplotlib kiwisolver typing extensions package ucrt conflicts for pytest cov coverage ucrt openblas runtime ucrt ucrt cython lint cython ucrt click ucrt setuptools ucrt pytest ucrt ruff ucrt pytest xdist psutil ucrt pkg config libglib ucrt mypy ucrt mpmath ucrt pythran numpy ucrt matplotlib matplotlib base ucrt ucrt ninja runtime ucrt cython ucrt python libsqlite ucrt numpy ucrt asv runtime ucrt package dataclasses conflicts for ipython black dataclasses rich click rich dataclasses package flit core conflicts for typing extensions flit core mypy typing extensions flit core package doit conflicts for doit pydevtool doit package pytz conflicts for sphinx babel pytz matplotlib pytz package requests conflicts for python pip requests sphinx requests sphinx design sphinx requests pooch requests pydata sphinx theme sphinx requests numpydoc sphinx requests package libblas conflicts for numpy libblas pythran numpy libblas numpy libcblas libblas matplotlib numpy libblas libblas package pygments conflicts for sphinx pygments rich click rich pygments pydata sphinx theme sphinx pygments ipython pygments numpydoc sphinx pygments sphinx design sphinx pygments package pyparsing conflicts for pooch packaging pyparsing matplotlib pyparsing pydata sphinx theme packaging pyparsing sphinx packaging pyparsing pytest packaging pyparsing matplotlib matplotlib base pyparsing package ninja conflicts for meson ninja meson python ninja meson python meson ninja ninja package decorator conflicts for ipython decorator pythran decorator package llvm openmp conflicts for openblas libflang llvm openmp numpy libflang llvm openmp package click conflicts for ipython black click rich click click click package tomlkit conflicts for meson python tomli tomlkit pytest tomli tomlkit package backports conflicts for pooch backports tempfile backports ipython backports shutil get terminal size backports matplotlib backports functools lru cache backports package sphinxcontrib websupport conflicts for sphinx sphinxcontrib websupport numpydoc sphinx sphinxcontrib websupport pydata sphinx theme sphinx sphinxcontrib websupport package scandir conflicts for pytest scandir ipython scandir package meson conflicts for meson meson python meson package wheel conflicts for python pip wheel meson python wheel package charset normalizer conflicts for sphinx requests charset normalizer pooch requests charset normalizer package attrs conflicts for pytest timeout pytest attrs pytest cov pytest attrs ipython black attrs pytest xdist pytest attrs pytest attrs package openblas conflicts for numpy libblas openblas openblas matplotlib numpy openblas numpy openblas libblas openblas package psutil conflicts for mypy psutil pytest xdist psutil package cython conflicts for cython cython lint cython package conflicts for pytest importlib metadata click importlib metadata following specifications were found to be incompatible with your system feature win win feature win win click unix click win ipython linux ipython win rich click click unix rich click click win setuptools wincertstore win your installed version is scipy numpy python version information python main aug anaconda inc on type help copyright credits or license for more information import scipy scipy show config lapack mkl info libraries library dirs define macros include dirs lapack opt info libraries library dirs define macros include dirs blas mkl info libraries library dirs define macros include dirs blas opt info libraries library dirs define macros include dirs import sys import numpy print scipy version numpy version sys version info sys version info major minor micro releaselevel final serial
| 1
|
171,321
| 27,099,222,064
|
IssuesEvent
|
2023-02-15 07:06:05
|
microsoft/WSA
|
https://api.github.com/repos/microsoft/WSA
|
closed
|
Want to turn off split screen mode
|
by-design fix-inbound feature-request
|
### Is your feature request related to a problem? Please describe
At present, WSA automatically turns on the split screen mode, which will bring inconvenience to the experience of some applications (including but not limited to: unable to open the scan QR code interface, and video cannot be played in full screen). I don't know your purpose of turning on the split screen mode, but from the result, it has brought me a bad experience. I hope to solve this problem in the subsequent version. Thank you
### Describe the solution you'd like
Turn off split screen mode by default or add a switch to let the user choose whether to turn on split screen mode
### Describe alternatives you've considered
_No response_
### Please specify the version of Windows Subsystem for Android
2301.40000.4.0
|
1.0
|
Want to turn off split screen mode - ### Is your feature request related to a problem? Please describe
At present, WSA automatically turns on the split screen mode, which will bring inconvenience to the experience of some applications (including but not limited to: unable to open the scan QR code interface, and video cannot be played in full screen). I don't know your purpose of turning on the split screen mode, but from the result, it has brought me a bad experience. I hope to solve this problem in the subsequent version. Thank you
### Describe the solution you'd like
Turn off split screen mode by default or add a switch to let the user choose whether to turn on split screen mode
### Describe alternatives you've considered
_No response_
### Please specify the version of Windows Subsystem for Android
2301.40000.4.0
|
non_defect
|
want to turn off split screen mode is your feature request related to a problem please describe at present wsa automatically turns on the split screen mode which will bring inconvenience to the experience of some applications including but not limited to unable to open the scan qr code interface and video cannot be played in full screen i don t know your purpose of turning on the split screen mode but from the result it has brought me a bad experience i hope to solve this problem in the subsequent version thank you describe the solution you d like turn off split screen mode by default or add a switch to let the user choose whether to turn on split screen mode describe alternatives you ve considered no response please specify the version of windows subsystem for android
| 0
|
42,330
| 10,965,856,149
|
IssuesEvent
|
2019-11-28 04:48:23
|
vector-im/riot-web
|
https://api.github.com/repos/vector-im/riot-web
|
closed
|
Message Editing: Remove button is available in edit history even if I don't have permissions to remove the edit
|
bug defect feature:aggregations feature:message-editing phase:3
|
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
The "Remove" button appears on message edits even if I don't have the PL to remove that message.
### Steps to reproduce
- Be in a room with someone
- Have them send a message and edit it
- See that you have a "Remove" button when viewing the edit history for that message
Riot should check if I'm able to redact the edit first before showing that button.
<!-- Please send us logs for your bug report. They're very important for bugs
which are hard to reproduce. To do this, create this issue then go to your
account settings and click 'Submit Debug Logs' from the Help & About tab -->
Logs being sent: no
<!-- Include screenshots if possible: you can drag and drop images below. -->
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web
- **Browser**: Firefox
- **OS**: Debian Stretch
- **URL**: riot.im/develop
|
1.0
|
Message Editing: Remove button is available in edit history even if I don't have permissions to remove the edit - <!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
The "Remove" button appears on message edits even if I don't have the PL to remove that message.
### Steps to reproduce
- Be in a room with someone
- Have them send a message and edit it
- See that you have a "Remove" button when viewing the edit history for that message
Riot should check if I'm able to redact the edit first before showing that button.
<!-- Please send us logs for your bug report. They're very important for bugs
which are hard to reproduce. To do this, create this issue then go to your
account settings and click 'Submit Debug Logs' from the Help & About tab -->
Logs being sent: no
<!-- Include screenshots if possible: you can drag and drop images below. -->
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web
- **Browser**: Firefox
- **OS**: Debian Stretch
- **URL**: riot.im/develop
|
defect
|
message editing remove button is available in edit history even if i don t have permissions to remove the edit this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description the remove button appears on message edits even if i don t have the pl to remove that message steps to reproduce be in a room with someone have them send a message and edit it see that you have a remove button when viewing the edit history for that message riot should check if i m able to redact the edit first before showing that button please send us logs for your bug report they re very important for bugs which are hard to reproduce to do this create this issue then go to your account settings and click submit debug logs from the help about tab logs being sent no version information platform web browser firefox os debian stretch url riot im develop
| 1
|
46,270
| 13,055,882,493
|
IssuesEvent
|
2020-07-30 03:00:54
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
genie-icetray - link against TTree (Trac #894)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
Migrated from https://code.icecube.wisc.edu/ticket/894
```json
{
"status": "closed",
"changetime": "2015-03-04T00:24:00",
"description": "It really helps when you specify what you need to link with, instead of assuming it magically happens. Here's a patch that works for me. No promises for anyone else.\n\n\n{{{\nIndex: ../../../../metaprojects/simulation/V04-01-11/genie-icetray/CMakeLists.txt\n===================================================================\n--- ../../../../metaprojects/simulation/V04-01-11/genie-icetray/CMakeLists.txt (revision 123305)\n+++ ../../../../metaprojects/simulation/V04-01-11/genie-icetray/CMakeLists.txt (working copy)\n@@ -37,6 +37,7 @@\n # we need three extra ROOT libs\n find_library(ROOT_GEOM_LIB Geom PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n find_library(ROOT_EG_LIB EG PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n+ find_library(ROOT_TREE_LIB Tree PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n find_library(ROOT_EGPYTHIA6_LIB EGPythia6 PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n find_library(LHAPDF_LIB LHAPDF PATHS ${I3_PORTS}/lib NO_DEFAULT_PATH)\n find_library(PYTHIA_SO_LIB Pythia6 PATHS ${I3_PORTS}/lib NO_DEFAULT_PATH)\n@@ -49,7 +50,7 @@\n USE_TOOLS python boost root log4cpp xml2 genie\n USE_PROJECTS icetray dataio dataclasses phys-services\n \n- LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n+ LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_TREE_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n )\n # SET_TARGET_PROPERTIES(genie-icetray PROPERTIES LINK_FLAGS -Wl,-u,pydata_)\n \n@@ -61,7 +62,7 @@\n USE_TOOLS python boost root log4cpp xml2 genie\n USE_PROJECTS icetray genie-icetray\n \n- LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n+ LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_TREE_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n )\n \n i3_executable(spladd\n@@ -69,7 +70,7 @@\n USE_TOOLS python boost root log4cpp xml2 genie\n USE_PROJECTS icetray genie-icetray\n \n- LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n+ LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_TREE_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n )\n endif (GENIE_ICETRAY_EXEC)\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1425428640827800",
"component": "combo simulation",
"summary": "genie-icetray - link against TTree",
"priority": "critical",
"keywords": "genie-icetray",
"time": "2015-03-04T00:07:08",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
1.0
|
genie-icetray - link against TTree (Trac #894) - Migrated from https://code.icecube.wisc.edu/ticket/894
```json
{
"status": "closed",
"changetime": "2015-03-04T00:24:00",
"description": "It really helps when you specify what you need to link with, instead of assuming it magically happens. Here's a patch that works for me. No promises for anyone else.\n\n\n{{{\nIndex: ../../../../metaprojects/simulation/V04-01-11/genie-icetray/CMakeLists.txt\n===================================================================\n--- ../../../../metaprojects/simulation/V04-01-11/genie-icetray/CMakeLists.txt (revision 123305)\n+++ ../../../../metaprojects/simulation/V04-01-11/genie-icetray/CMakeLists.txt (working copy)\n@@ -37,6 +37,7 @@\n # we need three extra ROOT libs\n find_library(ROOT_GEOM_LIB Geom PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n find_library(ROOT_EG_LIB EG PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n+ find_library(ROOT_TREE_LIB Tree PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n find_library(ROOT_EGPYTHIA6_LIB EGPythia6 PATHS ${ROOTSYS}/lib NO_DEFAULT_PATH)\n find_library(LHAPDF_LIB LHAPDF PATHS ${I3_PORTS}/lib NO_DEFAULT_PATH)\n find_library(PYTHIA_SO_LIB Pythia6 PATHS ${I3_PORTS}/lib NO_DEFAULT_PATH)\n@@ -49,7 +50,7 @@\n USE_TOOLS python boost root log4cpp xml2 genie\n USE_PROJECTS icetray dataio dataclasses phys-services\n \n- LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n+ LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_TREE_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n )\n # SET_TARGET_PROPERTIES(genie-icetray PROPERTIES LINK_FLAGS -Wl,-u,pydata_)\n \n@@ -61,7 +62,7 @@\n USE_TOOLS python boost root log4cpp xml2 genie\n USE_PROJECTS icetray genie-icetray\n \n- LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n+ LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_TREE_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n )\n \n i3_executable(spladd\n@@ -69,7 +70,7 @@\n USE_TOOLS python boost root log4cpp xml2 genie\n USE_PROJECTS icetray genie-icetray\n \n- LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n+ LINK_LIBRARIES ${PYTHIA_SO_LIB} ${ROOT_GEOM_LIB} ${ROOT_EG_LIB} ${ROOT_TREE_LIB} ${ROOT_EGPYTHIA6_LIB} ${LHAPDF_LIB}\n )\n endif (GENIE_ICETRAY_EXEC)\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1425428640827800",
"component": "combo simulation",
"summary": "genie-icetray - link against TTree",
"priority": "critical",
"keywords": "genie-icetray",
"time": "2015-03-04T00:07:08",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
|
defect
|
genie icetray link against ttree trac migrated from json status closed changetime description it really helps when you specify what you need to link with instead of assuming it magically happens here s a patch that works for me no promises for anyone else n n n nindex metaprojects simulation genie icetray cmakelists txt n n metaprojects simulation genie icetray cmakelists txt revision n metaprojects simulation genie icetray cmakelists txt working copy n n we need three extra root libs n find library root geom lib geom paths rootsys lib no default path n find library root eg lib eg paths rootsys lib no default path n find library root tree lib tree paths rootsys lib no default path n find library root lib paths rootsys lib no default path n find library lhapdf lib lhapdf paths ports lib no default path n find library pythia so lib paths ports lib no default path n n use tools python boost root genie n use projects icetray dataio dataclasses phys services n n link libraries pythia so lib root geom lib root eg lib root lib lhapdf lib n link libraries pythia so lib root geom lib root eg lib root tree lib root lib lhapdf lib n n set target properties genie icetray properties link flags wl u pydata n n n use tools python boost root genie n use projects icetray genie icetray n n link libraries pythia so lib root geom lib root eg lib root lib lhapdf lib n link libraries pythia so lib root geom lib root eg lib root tree lib root lib lhapdf lib n n n executable spladd n n use tools python boost root genie n use projects icetray genie icetray n n link libraries pythia so lib root geom lib root eg lib root lib lhapdf lib n link libraries pythia so lib root geom lib root eg lib root tree lib root lib lhapdf lib n n endif genie icetray exec n reporter david schultz cc resolution fixed ts component combo simulation summary genie icetray link against ttree priority critical keywords genie icetray time milestone owner olivas type defect
| 1
|
14,258
| 2,796,736,797
|
IssuesEvent
|
2015-05-12 09:30:04
|
marcopar/ethernetip4j
|
https://api.github.com/repos/marcopar/ethernetip4j
|
closed
|
Cumulative patch for various bugs
|
auto-migrated Priority-Medium Type-Defect
|
```
Sorry but my tree has diverged too much to be able to send single patched to
individual issues so i gathering everything here
- fix for segment lenght calculation in CipMessageRouterRequest that causes
padding errors and consecuent failure of the request (issue 3)
- fix for errors when accessing array positions larger than 255
- fix for swapped parameters routePathLinkedAddress and routePathPort in
CipCommandSpecificDataRequest.fillBuffer (issue 2)
- fix for array handling in SimpleLogixCommunicator (issue 2)
- better error handling with extended status
- string support in SimpleLogixCommunicator
```
Original issue reported on code.google.com by `marco...@gmail.com` on 27 Mar 2014 at 5:23
Attachments:
* [a.patch](https://storage.googleapis.com/google-code-attachments/ethernetip4j/issue-4/comment-0/a.patch)
|
1.0
|
Cumulative patch for various bugs - ```
Sorry but my tree has diverged too much to be able to send single patched to
individual issues so i gathering everything here
- fix for segment lenght calculation in CipMessageRouterRequest that causes
padding errors and consecuent failure of the request (issue 3)
- fix for errors when accessing array positions larger than 255
- fix for swapped parameters routePathLinkedAddress and routePathPort in
CipCommandSpecificDataRequest.fillBuffer (issue 2)
- fix for array handling in SimpleLogixCommunicator (issue 2)
- better error handling with extended status
- string support in SimpleLogixCommunicator
```
Original issue reported on code.google.com by `marco...@gmail.com` on 27 Mar 2014 at 5:23
Attachments:
* [a.patch](https://storage.googleapis.com/google-code-attachments/ethernetip4j/issue-4/comment-0/a.patch)
|
defect
|
cumulative patch for various bugs sorry but my tree has diverged too much to be able to send single patched to individual issues so i gathering everything here fix for segment lenght calculation in cipmessagerouterrequest that causes padding errors and consecuent failure of the request issue fix for errors when accessing array positions larger than fix for swapped parameters routepathlinkedaddress and routepathport in cipcommandspecificdatarequest fillbuffer issue fix for array handling in simplelogixcommunicator issue better error handling with extended status string support in simplelogixcommunicator original issue reported on code google com by marco gmail com on mar at attachments
| 1
|
120,214
| 12,061,448,974
|
IssuesEvent
|
2020-04-15 23:50:11
|
jooby-project/jooby
|
https://api.github.com/repos/jooby-project/jooby
|
closed
|
Jooby2 documentation: It'd be great to see example code for one Jooby instance handle two domains
|
documentation
|
Both on say 8080, but requests to http://foo.devd.io and http://bar.devd.io (both redirect to 127.0.0.1) serve different content.
Examples for the script and MVC modes of operation would be very cool :)
|
1.0
|
Jooby2 documentation: It'd be great to see example code for one Jooby instance handle two domains - Both on say 8080, but requests to http://foo.devd.io and http://bar.devd.io (both redirect to 127.0.0.1) serve different content.
Examples for the script and MVC modes of operation would be very cool :)
|
non_defect
|
documentation it d be great to see example code for one jooby instance handle two domains both on say but requests to and both redirect to serve different content examples for the script and mvc modes of operation would be very cool
| 0
|
21,123
| 3,461,696,193
|
IssuesEvent
|
2015-12-20 09:26:34
|
arti01/jkursy
|
https://api.github.com/repos/arti01/jkursy
|
closed
|
nowy użytkownik i przypisanie kursu
|
auto-migrated Priority-Low Type-Defect
|
```
Dodaję użytkownika z pozycji admina. Nie mogę od razu przypisać mu kursu,
muszę wysłać dane, wejść w usera jeszcze raz i dopiero wtedy przypisać mu
konkretny kurs. Dobrze by było zrobić to od razu.
```
Original issue reported on code.google.com by `juko...@gmail.com` on 10 Feb 2011 at 7:26
|
1.0
|
nowy użytkownik i przypisanie kursu - ```
Dodaję użytkownika z pozycji admina. Nie mogę od razu przypisać mu kursu,
muszę wysłać dane, wejść w usera jeszcze raz i dopiero wtedy przypisać mu
konkretny kurs. Dobrze by było zrobić to od razu.
```
Original issue reported on code.google.com by `juko...@gmail.com` on 10 Feb 2011 at 7:26
|
defect
|
nowy użytkownik i przypisanie kursu dodaję użytkownika z pozycji admina nie mogę od razu przypisać mu kursu muszę wysłać dane wejść w usera jeszcze raz i dopiero wtedy przypisać mu konkretny kurs dobrze by było zrobić to od razu original issue reported on code google com by juko gmail com on feb at
| 1
|
169,603
| 6,412,831,579
|
IssuesEvent
|
2017-08-08 05:18:14
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
closed
|
Support for GCS Cloud Pub/Sub Notifications
|
priority: p2+ type: enhancement
|
GCS has a new feature (currently in private alpha) wherein buckets can be configured to send object change notifications to Cloud Pub/Sub topics, rather than as push notifications to HTTPS endpoints.
For gcloud-python, I imagine such a feature might look like this:
from gcloud import pubsub
from gcloud import storage
storage_client = storage.Client()
pubsub_client = pubsub.Client()
bucket = storage_client.get_bucket(BUCKET_ID)
topic = pubsub_client.topic(TOPIC_ID)
notification_config = bucket.send_notifications_to(topic)
This feature was previously filed as https://github.com/GoogleCloudPlatform/google-cloud-python/issues/2768, but that one was closed because the feature was still in private alpha. However, today the feature is now in public beta.
|
1.0
|
Support for GCS Cloud Pub/Sub Notifications - GCS has a new feature (currently in private alpha) wherein buckets can be configured to send object change notifications to Cloud Pub/Sub topics, rather than as push notifications to HTTPS endpoints.
For gcloud-python, I imagine such a feature might look like this:
from gcloud import pubsub
from gcloud import storage
storage_client = storage.Client()
pubsub_client = pubsub.Client()
bucket = storage_client.get_bucket(BUCKET_ID)
topic = pubsub_client.topic(TOPIC_ID)
notification_config = bucket.send_notifications_to(topic)
This feature was previously filed as https://github.com/GoogleCloudPlatform/google-cloud-python/issues/2768, but that one was closed because the feature was still in private alpha. However, today the feature is now in public beta.
|
non_defect
|
support for gcs cloud pub sub notifications gcs has a new feature currently in private alpha wherein buckets can be configured to send object change notifications to cloud pub sub topics rather than as push notifications to https endpoints for gcloud python i imagine such a feature might look like this from gcloud import pubsub from gcloud import storage storage client storage client pubsub client pubsub client bucket storage client get bucket bucket id topic pubsub client topic topic id notification config bucket send notifications to topic this feature was previously filed as but that one was closed because the feature was still in private alpha however today the feature is now in public beta
| 0
|
216,509
| 16,768,104,592
|
IssuesEvent
|
2021-06-14 11:33:36
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[Event Filtering] Hide content while checking if entries exists on event filters list
|
OLM Sprint QA:Ready for Testing Team:Onboarding and Lifecycle Mgt bug v7.14.0
|
Content is displayed while checking if entries exists. This can cause weird behaviours (flash) when there is no data or there are existing entries.
**Steps to reproduce:**
1. Go to Administration tab
2. Go to Event Filters tab
3. When no entries exists there is a strange behaviour.
A loader should appear first time when discovering if there are entries or entries exists.

|
1.0
|
[Event Filtering] Hide content while checking if entries exists on event filters list - Content is displayed while checking if entries exists. This can cause weird behaviours (flash) when there is no data or there are existing entries.
**Steps to reproduce:**
1. Go to Administration tab
2. Go to Event Filters tab
3. When no entries exists there is a strange behaviour.
A loader should appear first time when discovering if there are entries or entries exists.

|
non_defect
|
hide content while checking if entries exists on event filters list content is displayed while checking if entries exists this can cause weird behaviours flash when there is no data or there are existing entries steps to reproduce go to administration tab go to event filters tab when no entries exists there is a strange behaviour a loader should appear first time when discovering if there are entries or entries exists
| 0
|
532,624
| 15,560,404,027
|
IssuesEvent
|
2021-03-16 12:40:44
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
USER ISSUE: Meteor Blowing up the World gives negative numbers for All types of Trees
|
Category: Gameplay Priority: Medium Squad: Otter Status: Not reproduced Type: Bug
|
**Version:** 0.7.7.2 beta

|
1.0
|
USER ISSUE: Meteor Blowing up the World gives negative numbers for All types of Trees - **Version:** 0.7.7.2 beta

|
non_defect
|
user issue meteor blowing up the world gives negative numbers for all types of trees version beta
| 0
|
30,741
| 4,213,168,271
|
IssuesEvent
|
2016-06-29 18:15:36
|
phetsims/pendulum-lab
|
https://api.github.com/repos/phetsims/pendulum-lab
|
closed
|
Scaling of pendulum's length does not match design document.
|
design:general
|
According to the design document the version dev.9 of the Pendulum lab does not satisfy this mock-up.
**Current Version**

**Mock-up**

This will require scaling of the ruler as well.
|
1.0
|
Scaling of pendulum's length does not match design document. - According to the design document the version dev.9 of the Pendulum lab does not satisfy this mock-up.
**Current Version**

**Mock-up**

This will require scaling of the ruler as well.
|
non_defect
|
scaling of pendulum s length does not match design document according to the design document the version dev of the pendulum lab does not satisfy this mock up current version mock up this will require scaling of the ruler as well
| 0
|
15,773
| 2,869,065,410
|
IssuesEvent
|
2015-06-05 23:03:09
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Get error "Uncaught InvalidStateError: Failed to call 'register' on 'Document' for type 'polymer-element'
|
Area-Pkg NeedsInfo Pkg-Polymer PolymerMilestone-Next Priority-Medium Type-Defect
|
When I use a debug build of dartium and try to use the Dartium Observatory
I get the following error:
[24956:24956:0116/153752:ERROR:CONSOLE(5777)] "Uncaught InvalidStateError: Failed to call 'register' on 'Document' for type 'polymer-element': a type with that name is already registered.", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (5777)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Got reply 0 {"type":"IsolateList","members":[{"type":"Isolate","id":"isolates/7117","name":"dom-modify.dart$main-7117"}]}", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Completing 0", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Uncaught Error: NullError: Cannot call "goc" on null", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Stack Trace:
TypeError: Cannot call method 'goc' of undefined
at dart.J.DA (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:22450:33)
at Wf.gNy (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:11631:38)
at Wf.gYK (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:11566:11)
at dart.y1 (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:21119:6)
at dart.y1 (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:21121:38)
at z6.tI (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20433:6)
at ek.Qh (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20580:5)
at ek.dE.yc (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20456:6)
at Ed.xn (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20465:19)
at Ed.a0.qv (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:21259:29)
", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
|
1.0
|
Get error "Uncaught InvalidStateError: Failed to call 'register' on 'Document' for type 'polymer-element' - When I use a debug build of dartium and try to use the Dartium Observatory
I get the following error:
[24956:24956:0116/153752:ERROR:CONSOLE(5777)] "Uncaught InvalidStateError: Failed to call 'register' on 'Document' for type 'polymer-element': a type with that name is already registered.", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (5777)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Got reply 0 {"type":"IsolateList","members":[{"type":"Isolate","id":"isolates/7117","name":"dom-modify.dart$main-7117"}]}", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Completing 0", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Uncaught Error: NullError: Cannot call "goc" on null", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
[24956:24956:0116/153754:INFO:CONSOLE(17223)] "Stack Trace:
TypeError: Cannot call method 'goc' of undefined
at dart.J.DA (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:22450:33)
at Wf.gNy (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:11631:38)
at Wf.gYK (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:11566:11)
at dart.y1 (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:21119:6)
at dart.y1 (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:21121:38)
at z6.tI (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20433:6)
at ek.Qh (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20580:5)
at ek.dE.yc (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20456:6)
at Ed.xn (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:20465:19)
at Ed.a0.qv (chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js:21259:29)
", source: chrome-devtools://devtools/bundled/Observatory/index_devtools.html_bootstrap.dart.js (17223)
|
defect
|
get error uncaught invalidstateerror failed to call register on document for type polymer element when i use a debug build of dartium and try to use the dartium observatory i get the following error quot uncaught invalidstateerror failed to call register on document for type polymer element a type with that name is already registered quot source chrome devtools devtools bundled observatory index devtools html bootstrap dart js quot got reply quot type quot quot isolatelist quot quot members quot quot source chrome devtools devtools bundled observatory index devtools html bootstrap dart js quot completing quot source chrome devtools devtools bundled observatory index devtools html bootstrap dart js quot uncaught error nullerror cannot call quot goc quot on null quot source chrome devtools devtools bundled observatory index devtools html bootstrap dart js quot stack trace typeerror cannot call method goc of undefined nbsp nbsp nbsp nbsp at dart j da chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at wf gny chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at wf gyk chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at dart chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at dart chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at ti chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at ek qh chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at ek de yc chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at ed xn chrome devtools devtools bundled observatory index devtools html bootstrap dart js nbsp nbsp nbsp nbsp at ed qv chrome devtools devtools bundled observatory index devtools html bootstrap dart js quot source chrome devtools devtools bundled observatory index devtools html bootstrap dart js
| 1
|
252,151
| 27,230,329,035
|
IssuesEvent
|
2023-02-21 12:47:19
|
timf-app-demo/Java-Demo
|
https://api.github.com/repos/timf-app-demo/Java-Demo
|
opened
|
CVE-2023-24998 (Medium) detected in commons-fileupload-1.3.1.jar
|
security vulnerability
|
## CVE-2023-24998 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-fileupload-1.3.1.jar</b></p></summary>
<p>The Apache Commons FileUpload component provides a simple yet flexible means of adding support for multipart
file upload functionality to servlets and web applications.</p>
<p>Library home page: <a href="http://commons.apache.org/proper/commons-fileupload/">http://commons.apache.org/proper/commons-fileupload/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-fileupload/commons-fileupload/1.3.1/commons-fileupload-1.3.1.jar</p>
<p>
Dependency Hierarchy:
- esapi-2.1.0.1.jar (Root Library)
- :x: **commons-fileupload-1.3.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/timf-app-demo/Java-Demo/commit/03a2d50646afdd26f30b31b1f14f22aab9de8e02">03a2d50646afdd26f30b31b1f14f22aab9de8e02</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons FileUpload before 1.5 does not limit the number of request parts to be processed resulting in the possibility of an attacker triggering a DoS with a malicious upload or series of uploads.
<p>Publish Date: 2023-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-24998>CVE-2023-24998</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://seclists.org/oss-sec/2023/q1/108">https://seclists.org/oss-sec/2023/q1/108</a></p>
<p>Release Date: 2023-02-20</p>
<p>Fix Resolution: commons-fileupload:commons-fileupload:1.5;org.apache.tomcat:tomcat-coyote:8.5.85,9.0.71,10.1.5</p>
</p>
</details>
<p></p>
|
True
|
CVE-2023-24998 (Medium) detected in commons-fileupload-1.3.1.jar - ## CVE-2023-24998 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-fileupload-1.3.1.jar</b></p></summary>
<p>The Apache Commons FileUpload component provides a simple yet flexible means of adding support for multipart
file upload functionality to servlets and web applications.</p>
<p>Library home page: <a href="http://commons.apache.org/proper/commons-fileupload/">http://commons.apache.org/proper/commons-fileupload/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-fileupload/commons-fileupload/1.3.1/commons-fileupload-1.3.1.jar</p>
<p>
Dependency Hierarchy:
- esapi-2.1.0.1.jar (Root Library)
- :x: **commons-fileupload-1.3.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/timf-app-demo/Java-Demo/commit/03a2d50646afdd26f30b31b1f14f22aab9de8e02">03a2d50646afdd26f30b31b1f14f22aab9de8e02</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons FileUpload before 1.5 does not limit the number of request parts to be processed resulting in the possibility of an attacker triggering a DoS with a malicious upload or series of uploads.
<p>Publish Date: 2023-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-24998>CVE-2023-24998</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://seclists.org/oss-sec/2023/q1/108">https://seclists.org/oss-sec/2023/q1/108</a></p>
<p>Release Date: 2023-02-20</p>
<p>Fix Resolution: commons-fileupload:commons-fileupload:1.5;org.apache.tomcat:tomcat-coyote:8.5.85,9.0.71,10.1.5</p>
</p>
</details>
<p></p>
|
non_defect
|
cve medium detected in commons fileupload jar cve medium severity vulnerability vulnerable library commons fileupload jar the apache commons fileupload component provides a simple yet flexible means of adding support for multipart file upload functionality to servlets and web applications library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository commons fileupload commons fileupload commons fileupload jar dependency hierarchy esapi jar root library x commons fileupload jar vulnerable library found in head commit a href found in base branch main vulnerability details apache commons fileupload before does not limit the number of request parts to be processed resulting in the possibility of an attacker triggering a dos with a malicious upload or series of uploads publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons fileupload commons fileupload org apache tomcat tomcat coyote
| 0
|
283,855
| 21,335,284,454
|
IssuesEvent
|
2022-04-18 13:54:18
|
francescaseverino/Focus-Bot
|
https://api.github.com/repos/francescaseverino/Focus-Bot
|
opened
|
Add meaningful comments
|
documentation
|
add comments to describe certain functions so that it is better understood by developers
|
1.0
|
Add meaningful comments - add comments to describe certain functions so that it is better understood by developers
|
non_defect
|
add meaningful comments add comments to describe certain functions so that it is better understood by developers
| 0
|
677,804
| 23,176,591,501
|
IssuesEvent
|
2022-07-31 14:07:13
|
prologin/sadm
|
https://api.github.com/repos/prologin/sadm
|
closed
|
workernode should get their tasks from the heartbeat
|
type:enhancement priority:major state:new
|
> ###### Originally posted 2015-11-11 17:25 by [serialk](https://bitbucket.org/serialk/) on Bitbucket
When using the master/worker system on a network we don't control, an important number of problems arise, especially when some of the workers are behind a NAT.
Currently, the system works this way:
- The workernode sends a heartbeat to the masternode
- The masternode registers the workernode and its hostname if he does not know it yet
- When a task has to be dispatched, the masternode does an RPC request to the worker to send the task
- When the task is complete, the workernode does an RPC request to the master to send the result back
This requires *all* the workers to be accessible from the masternode. This prevents us from running workers behind a NAT.
A better way to do that would be to send the tasks as a response to the heartbeat request. When a workernode sends a heartbeat, we check if it has any new tasks attributed, and we reply to the heartbeat request with these new tasks.
This will require a lot of changes in the code. What do you think of this design change?
|
1.0
|
workernode should get their tasks from the heartbeat - > ###### Originally posted 2015-11-11 17:25 by [serialk](https://bitbucket.org/serialk/) on Bitbucket
When using the master/worker system on a network we don't control, an important number of problems arise, especially when some of the workers are behind a NAT.
Currently, the system works this way:
- The workernode sends a heartbeat to the masternode
- The masternode registers the workernode and its hostname if he does not know it yet
- When a task has to be dispatched, the masternode does an RPC request to the worker to send the task
- When the task is complete, the workernode does an RPC request to the master to send the result back
This requires *all* the workers to be accessible from the masternode. This prevents us from running workers behind a NAT.
A better way to do that would be to send the tasks as a response to the heartbeat request. When a workernode sends a heartbeat, we check if it has any new tasks attributed, and we reply to the heartbeat request with these new tasks.
This will require a lot of changes in the code. What do you think of this design change?
|
non_defect
|
workernode should get their tasks from the heartbeat originally posted by on bitbucket when using the master worker system on a network we don t control an important number of problems arise especially when some of the workers are behind a nat currently the system works this way the workernode sends a heartbeat to the masternode the masternode registers the workernode and its hostname if he does not know it yet when a task has to be dispatched the masternode does an rpc request to the worker to send the task when the task is complete the workernode does an rpc request to the master to send the result back this requires all the workers to be accessible from the masternode this prevents us from running workers behind a nat a better way to do that would be to send the tasks as a response to the heartbeat request when a workernode sends a heartbeat we check if it has any new tasks attributed and we reply to the heartbeat request with these new tasks this will require a lot of changes in the code what do you think of this design change
| 0
|
61,964
| 17,023,820,438
|
IssuesEvent
|
2021-07-03 04:01:37
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
remove trailing whitespace in tags
|
Component: potlatch2 Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 5.52pm, Thursday, 6th September 2012]**
Potlatch2 doesn't remove trailing whitespace in tags.
For example [http://www.openstreetmap.org/api/0.6/node/1864014640/2 v2 of node 1864014640] has a broken name tag:
```
<tag k="name " v="Hopfgarten"/>
```
I can't see any point in keeping trailing whitespace, it should be discarded automatically.
josm removes trailing whitespace for keys and values.
|
1.0
|
remove trailing whitespace in tags - **[Submitted to the original trac issue database at 5.52pm, Thursday, 6th September 2012]**
Potlatch2 doesn't remove trailing whitespace in tags.
For example [http://www.openstreetmap.org/api/0.6/node/1864014640/2 v2 of node 1864014640] has a broken name tag:
```
<tag k="name " v="Hopfgarten"/>
```
I can't see any point in keeping trailing whitespace, it should be discarded automatically.
josm removes trailing whitespace for keys and values.
|
defect
|
remove trailing whitespace in tags doesn t remove trailing whitespace in tags for example has a broken name tag i can t see any point in keeping trailing whitespace it should be discarded automatically josm removes trailing whitespace for keys and values
| 1
|
8,099
| 2,611,452,255
|
IssuesEvent
|
2015-02-27 05:00:05
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
Random mines are spawned inside the ground
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Generate maps with lot of random mines.
2. Play on it.
What is the expected output? What do you see instead?
It doesn't happen often, but sometimes the mines are spawned inside the ground.
What version of the product are you using? On what operating system?
0.9.13 on Windows XP SP 2
Please provide any additional information below.
Screenshot attached
```
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 1 Oct 2010 at 4:02
Attachments:
* [Mine Bug.bmp](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-78/comment-0/Mine Bug.bmp)
|
1.0
|
Random mines are spawned inside the ground - ```
What steps will reproduce the problem?
1. Generate maps with lot of random mines.
2. Play on it.
What is the expected output? What do you see instead?
It doesn't happen often, but sometimes the mines are spawned inside the ground.
What version of the product are you using? On what operating system?
0.9.13 on Windows XP SP 2
Please provide any additional information below.
Screenshot attached
```
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 1 Oct 2010 at 4:02
Attachments:
* [Mine Bug.bmp](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-78/comment-0/Mine Bug.bmp)
|
defect
|
random mines are spawned inside the ground what steps will reproduce the problem generate maps with lot of random mines play on it what is the expected output what do you see instead it doesn t happen often but sometimes the mines are spawned inside the ground what version of the product are you using on what operating system on windows xp sp please provide any additional information below screenshot attached original issue reported on code google com by adibiaz gmail com on oct at attachments bug bmp
| 1
|
434,912
| 30,475,875,005
|
IssuesEvent
|
2023-07-17 16:27:55
|
cloudflare/cloudflare-docs
|
https://api.github.com/repos/cloudflare/cloudflare-docs
|
closed
|
Add programmatic verification code for Javascript
|
documentation content:new
|
### Which Cloudflare product(s) does this pertain to?
Zero Trust
### Subject Matter
Add programmatic verification code for Javascript
### Content Location
https://developers.cloudflare.com/cloudflare-one/identity/authorization-cookie/validating-json/#programmatic-verification
### Additional information
I am happy to pick this up
|
1.0
|
Add programmatic verification code for Javascript - ### Which Cloudflare product(s) does this pertain to?
Zero Trust
### Subject Matter
Add programmatic verification code for Javascript
### Content Location
https://developers.cloudflare.com/cloudflare-one/identity/authorization-cookie/validating-json/#programmatic-verification
### Additional information
I am happy to pick this up
|
non_defect
|
add programmatic verification code for javascript which cloudflare product s does this pertain to zero trust subject matter add programmatic verification code for javascript content location additional information i am happy to pick this up
| 0
|
824,454
| 31,157,294,301
|
IssuesEvent
|
2023-08-16 13:53:23
|
testomatio/app
|
https://api.github.com/repos/testomatio/app
|
closed
|
PROD: public report does not create at the first attempt
|
bug reporting ui\ux users priority medium
|
**Describe the bug**
Public report will be created only after page refresh
**To Reproduce**
Steps to reproduce the behavior:
1. go to Runs page
2. open a run report
3. click on menu button
4. click on publish
5. see the issue
**Screenshots**

**Desktop (please complete the following information):**
- OS: MacOS
- Browser chrome
- Application: production
|
1.0
|
PROD: public report does not create at the first attempt - **Describe the bug**
Public report will be created only after page refresh
**To Reproduce**
Steps to reproduce the behavior:
1. go to Runs page
2. open a run report
3. click on menu button
4. click on publish
5. see the issue
**Screenshots**

**Desktop (please complete the following information):**
- OS: MacOS
- Browser chrome
- Application: production
|
non_defect
|
prod public report does not create at the first attempt describe the bug public report will be created only after page refresh to reproduce steps to reproduce the behavior go to runs page open a run report click on menu button click on publish see the issue screenshots desktop please complete the following information os macos browser chrome application production
| 0
|
46,892
| 13,055,998,279
|
IssuesEvent
|
2020-07-30 03:20:48
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
Large objects are serialized incorrectly (Trac #2096)
|
Incomplete Migration Migrated from Trac combo core defect
|
Migrated from https://code.icecube.wisc.edu/ticket/2096
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:49",
"description": "It seems that at someone decided to store collection sizes in our `portable_binary_archive`s [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/serialization/trunk/public/archive/portable_binary_archive.hpp#L73 as 32 bits]. While it might seem unlikely that we would ever need to serialize an array of 2^32^ objects, this can happen because `I3Frame::save` [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/icetray/trunk/private/icetray/I3Frame.cxx#L566 serializes all frame objects to binary blobs], stored as `std::vector<char>`. Thus, any frame object whose serialized size exceeds 4GB has an incorrect size written (although its contents are still written out in full). The error becomes obvious on reading, when the implementation tries to read only the incorrectly reduced length, which usually causes the next read operation to fail (due to it unexpectedly occurring in the middle of the data of the very long object which hasn't been completely read) even before it can be discovered that the data which was read is incomplete (which would likely be detected by the checksumming). This is demonstrated fairly minimally by the following test case:\n\n{{{\n#include <iostream>\n#include <fstream>\n#include <vector>\n#include <icetray/serialization.h>\n\nint main(){\n\t{ //write\n\t\tstd::vector<char> v((1ULL<<32) + 100,'a');\n\t\tstd::cout << \"Writing vector with length \" << v.size() << std::endl;\n\t\tstd::ofstream outfile(\"test.dat\",std::ios::out|std::ios::binary);\n\t\ticecube::archive::portable_binary_oarchive poa(outfile);\n\t\tpoa << v;\n\t}\n\t{ //read\n\t\tstd::vector<char> v;\n\t\tstd::ifstream infile(\"test.dat\",std::ios::in|std::ios::binary);\n\t\ticecube::archive::portable_binary_iarchive poa(infile);\n\t\tpoa >> v;\n\t\tstd::cout << \"Read vector with length \" << v.size() << std::endl;\n\t}\n}\n}}}\n\nwhich can be compiled using \n\n{{{\n$CXX `icetray-config icetray`\n}}}\n\nThe output is \n\n{{{\nWriting vector with length 4294967396\nRead vector with length 100\n}}}\n\nwhich is clearly wrong. \n\nIn order to fix this without breaking backwards compatibility, it seems that some fancy footwork will be required for `serialization::collection_size_type`, since we must continue to read 32 bit collection sizes (which are mostly correctly anyway). We should be able to replace the current, simple `sv_override` and `ld_override`, with something more clever which writes 32 bits by default, if the value being written fits in them, and otherwise writes a special value (2^32^-1?) and then records a 64 bit length, or similar. There is a minor issue that there is no length value which is technically `clean` for use to indicate overflow, since a file could validly contain any 32 bit pattern in that place, but it is genuinely unlikely that any file written so far contains an object of exactly the maximum allowed length. \n\nOld files containing large objects will be unrecoverable, as they are already corrupt, but it should be possible to support both old files with small objects (i.e. virtually all existing files) and be able to write and read large objects going forward. \n\nIt should also be noted that we make size assumptions in several other places, such as the `save_override`s for string types, and all basic integer tags used by the serialization library itself: `version_type`, `class_id_type`, etc. Even if we could just make all of these things 64 bits, we clearly don't want to for size reasons. ",
"reporter": "cweaver",
"cc": "cweaver",
"resolution": "fixed",
"_ts": "1550067289977126",
"component": "combo core",
"summary": "Large objects are serialized incorrectly",
"priority": "major",
"keywords": "",
"time": "2017-09-23T00:58:47",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
1.0
|
Large objects are serialized incorrectly (Trac #2096) - Migrated from https://code.icecube.wisc.edu/ticket/2096
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:49",
"description": "It seems that at someone decided to store collection sizes in our `portable_binary_archive`s [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/serialization/trunk/public/archive/portable_binary_archive.hpp#L73 as 32 bits]. While it might seem unlikely that we would ever need to serialize an array of 2^32^ objects, this can happen because `I3Frame::save` [http://code.icecube.wisc.edu/projects/icecube/browser/IceTray/projects/icetray/trunk/private/icetray/I3Frame.cxx#L566 serializes all frame objects to binary blobs], stored as `std::vector<char>`. Thus, any frame object whose serialized size exceeds 4GB has an incorrect size written (although its contents are still written out in full). The error becomes obvious on reading, when the implementation tries to read only the incorrectly reduced length, which usually causes the next read operation to fail (due to it unexpectedly occurring in the middle of the data of the very long object which hasn't been completely read) even before it can be discovered that the data which was read is incomplete (which would likely be detected by the checksumming). This is demonstrated fairly minimally by the following test case:\n\n{{{\n#include <iostream>\n#include <fstream>\n#include <vector>\n#include <icetray/serialization.h>\n\nint main(){\n\t{ //write\n\t\tstd::vector<char> v((1ULL<<32) + 100,'a');\n\t\tstd::cout << \"Writing vector with length \" << v.size() << std::endl;\n\t\tstd::ofstream outfile(\"test.dat\",std::ios::out|std::ios::binary);\n\t\ticecube::archive::portable_binary_oarchive poa(outfile);\n\t\tpoa << v;\n\t}\n\t{ //read\n\t\tstd::vector<char> v;\n\t\tstd::ifstream infile(\"test.dat\",std::ios::in|std::ios::binary);\n\t\ticecube::archive::portable_binary_iarchive poa(infile);\n\t\tpoa >> v;\n\t\tstd::cout << \"Read vector with length \" << v.size() << std::endl;\n\t}\n}\n}}}\n\nwhich can be compiled using \n\n{{{\n$CXX `icetray-config icetray`\n}}}\n\nThe output is \n\n{{{\nWriting vector with length 4294967396\nRead vector with length 100\n}}}\n\nwhich is clearly wrong. \n\nIn order to fix this without breaking backwards compatibility, it seems that some fancy footwork will be required for `serialization::collection_size_type`, since we must continue to read 32 bit collection sizes (which are mostly correctly anyway). We should be able to replace the current, simple `sv_override` and `ld_override`, with something more clever which writes 32 bits by default, if the value being written fits in them, and otherwise writes a special value (2^32^-1?) and then records a 64 bit length, or similar. There is a minor issue that there is no length value which is technically `clean` for use to indicate overflow, since a file could validly contain any 32 bit pattern in that place, but it is genuinely unlikely that any file written so far contains an object of exactly the maximum allowed length. \n\nOld files containing large objects will be unrecoverable, as they are already corrupt, but it should be possible to support both old files with small objects (i.e. virtually all existing files) and be able to write and read large objects going forward. \n\nIt should also be noted that we make size assumptions in several other places, such as the `save_override`s for string types, and all basic integer tags used by the serialization library itself: `version_type`, `class_id_type`, etc. Even if we could just make all of these things 64 bits, we clearly don't want to for size reasons. ",
"reporter": "cweaver",
"cc": "cweaver",
"resolution": "fixed",
"_ts": "1550067289977126",
"component": "combo core",
"summary": "Large objects are serialized incorrectly",
"priority": "major",
"keywords": "",
"time": "2017-09-23T00:58:47",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
defect
|
large objects are serialized incorrectly trac migrated from json status closed changetime description it seems that at someone decided to store collection sizes in our portable binary archive s while it might seem unlikely that we would ever need to serialize an array of objects this can happen because save stored as std vector thus any frame object whose serialized size exceeds has an incorrect size written although its contents are still written out in full the error becomes obvious on reading when the implementation tries to read only the incorrectly reduced length which usually causes the next read operation to fail due to it unexpectedly occurring in the middle of the data of the very long object which hasn t been completely read even before it can be discovered that the data which was read is incomplete which would likely be detected by the checksumming this is demonstrated fairly minimally by the following test case n n n include n include n include n include n nint main n t write n t tstd vector v v n t tstd ifstream infile test dat std ios in std ios binary n t ticecube archive portable binary iarchive poa infile n t tpoa v n t tstd cout read vector with length v size std endl n t n n n nwhich can be compiled using n n n cxx icetray config icetray n n nthe output is n n nwriting vector with length nread vector with length n n nwhich is clearly wrong n nin order to fix this without breaking backwards compatibility it seems that some fancy footwork will be required for serialization collection size type since we must continue to read bit collection sizes which are mostly correctly anyway we should be able to replace the current simple sv override and ld override with something more clever which writes bits by default if the value being written fits in them and otherwise writes a special value and then records a bit length or similar there is a minor issue that there is no length value which is technically clean for use to indicate overflow since a file could validly contain any bit pattern in that place but it is genuinely unlikely that any file written so far contains an object of exactly the maximum allowed length n nold files containing large objects will be unrecoverable as they are already corrupt but it should be possible to support both old files with small objects i e virtually all existing files and be able to write and read large objects going forward n nit should also be noted that we make size assumptions in several other places such as the save override s for string types and all basic integer tags used by the serialization library itself version type class id type etc even if we could just make all of these things bits we clearly don t want to for size reasons reporter cweaver cc cweaver resolution fixed ts component combo core summary large objects are serialized incorrectly priority major keywords time milestone owner david schultz type defect
| 1
|
356,183
| 10,589,720,822
|
IssuesEvent
|
2019-10-09 07:01:49
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
free.facebook.com - see bug description
|
browser-firefox-mobile engine-gecko priority-critical
|
<!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 6.0.1; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://free.facebook.com/aleahh.galakas?_rdc=2&_rdr
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 6.0.1
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: Unauthorized
**Steps to Reproduce**:
Unauthorized sell of drugs and guns remove this profile account
https://www.facebook.com/aleahh.galakas
[](https://webcompat.com/uploads/2019/10/8c900bb3-a88c-48a0-8a1a-a650107ef058.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191005090031</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
Submitted in the name of `@Annaros`
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
free.facebook.com - see bug description - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 6.0.1; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://free.facebook.com/aleahh.galakas?_rdc=2&_rdr
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 6.0.1
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: Unauthorized
**Steps to Reproduce**:
Unauthorized sell of drugs and guns remove this profile account
https://www.facebook.com/aleahh.galakas
[](https://webcompat.com/uploads/2019/10/8c900bb3-a88c-48a0-8a1a-a650107ef058.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191005090031</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
Submitted in the name of `@Annaros`
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
free facebook com see bug description url browser version firefox mobile operating system android tested another browser unknown problem type something else description unauthorized steps to reproduce unauthorized sell of drugs and guns remove this profile account browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false submitted in the name of annaros from with ❤️
| 0
|
388,160
| 11,483,933,867
|
IssuesEvent
|
2020-02-11 01:38:08
|
BendroCorp/bendrocorp-api
|
https://api.github.com/repos/BendroCorp/bendrocorp-api
|
closed
|
Events with zero auto attendance cannot be certified
|
bug priority: high
|
Events with zero auto attendance cannot be certified. The certification list comes back empty from the API.
|
1.0
|
Events with zero auto attendance cannot be certified - Events with zero auto attendance cannot be certified. The certification list comes back empty from the API.
|
non_defect
|
events with zero auto attendance cannot be certified events with zero auto attendance cannot be certified the certification list comes back empty from the api
| 0
|
77,624
| 27,081,800,027
|
IssuesEvent
|
2023-02-14 14:30:15
|
zed-industries/community
|
https://api.github.com/repos/zed-industries/community
|
closed
|
Toggle Side Bar with Key Binding cmd-shift-b Does Not Work
|
defect triage
|
### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
### Expected
When I use the key binding cmd-shift-b then the right sidebar (dock) opens or closes. As indicated in the menu and in the default key bindings:
` "cmd-shift-b": "workspace::ToggleRightSidebar"`
### Actual Behaviour
Given the right sidebar (dock):
- is closed, when I press cmd-shift-b then nothing happens, the sidebar remains closed
- is open, when I press cmd-shift-b then the sidebar closes
### Environment
Zed: v0.72.5 (stable)
OS: macOS 13.2.0
Memory: 16 GiB
Architecture: aarch64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
_No response_
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_
|
1.0
|
Toggle Side Bar with Key Binding cmd-shift-b Does Not Work - ### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
### Expected
When I use the key binding cmd-shift-b then the right sidebar (dock) opens or closes. As indicated in the menu and in the default key bindings:
` "cmd-shift-b": "workspace::ToggleRightSidebar"`
### Actual Behaviour
Given the right sidebar (dock):
- is closed, when I press cmd-shift-b then nothing happens, the sidebar remains closed
- is open, when I press cmd-shift-b then the sidebar closes
### Environment
Zed: v0.72.5 (stable)
OS: macOS 13.2.0
Memory: 16 GiB
Architecture: aarch64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
_No response_
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_
|
defect
|
toggle side bar with key binding cmd shift b does not work check for existing issues completed describe the bug provide steps to reproduce it expected when i use the key binding cmd shift b then the right sidebar dock opens or closes as indicated in the menu and in the default key bindings cmd shift b workspace togglerightsidebar actual behaviour given the right sidebar dock is closed when i press cmd shift b then nothing happens the sidebar remains closed is open when i press cmd shift b then the sidebar closes environment zed stable os macos memory gib architecture if applicable add mockups screenshots to help explain present your vision of the feature no response if applicable attach your library logs zed zed log file to this issue if you only need the most recent lines you can run the zed open log command palette action to see the last no response
| 1
|
217,544
| 16,855,806,216
|
IssuesEvent
|
2021-06-21 06:27:58
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
raftstore::test_merge::test_merge_isolated_store_with_no_target_peer failed
|
component/test
|
Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/39677/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/39677/display/redirect</a>
|
1.0
|
raftstore::test_merge::test_merge_isolated_store_with_no_target_peer failed - Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/39677/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tikv_ghpr_test/39677/display/redirect</a>
|
non_defect
|
raftstore test merge test merge isolated store with no target peer failed latest build a href
| 0
|
700,216
| 24,051,634,734
|
IssuesEvent
|
2022-09-16 13:19:17
|
unfoldingWord/translationCore
|
https://api.github.com/repos/unfoldingWord/translationCore
|
closed
|
On tC startup remove bibles processed with different usfm-js version update with bundled bibles
|
QA/Pass Priority/High Regression 3.3.0
|
## Story Explanation
### User Story
<!-- This is a short, simple description of the feature told from
the perspective of the person who desires the new capability -->
We have made a fix to USFM-js to preserve space before punctuation. This impacts the UHB and maybe other languages. Since the UHB is not changing that frequently, we run the risk that the user already has downloaded the latest UHB with tC 3.2.0, and then when they upgrade to 3.2.1 (for example), it will not replace the installed UHB with the bundled version because the version numbers are the same.
A few options:
1. On D43-Catalog and unfoldingWord, we increment the version number of UHB after the new tC is released - this could still be problematic
1. in the newer tCore we could add a special case for the UHB that we force update of UHB after tC upgrade
1. add processing date to the installed bibles so that we can compare to processing date of bundled resources to determine if we need to upgrade a bible
1. on tC version update we remove all the installed bibles and auto install bundled bibles (this is what we already do for tHelps). May be overkill, but safe for users, and less work for developers
1. if tC bundled resource is newer than installed resources we remove all the installed bibles and auto install bundled bibles (this is also what we already do for tHelps). May be overkill, but safe for users, and less work for developers. This is better for developers and QA since we don't have to worry about manually deleting resources before testing.
Number 5 is my recommendation. Has advantage of most safety going forward. But slows down first installs, and users will have to download again any custom bibles. 4 and 5 are probably the least effort since we already do this for tHelps.
### Features / Specifications
<!-- What are the high level features that need to be completed? -->
- [ ]
- [ ]
- [ ]
### Definition of Done
<!-- This is all the conditions, or acceptance criteria, that a
software product must satisfy are met and ready to be
accepted by a user, customer, team, or consuming system. -->
- [ ]
- [ ]
- [ ]
<!--
## QA
### Fails
- [ ]
- [ ]
-->
## Additional Context
### Mockups
|
1.0
|
On tC startup remove bibles processed with different usfm-js version update with bundled bibles - ## Story Explanation
### User Story
<!-- This is a short, simple description of the feature told from
the perspective of the person who desires the new capability -->
We have made a fix to USFM-js to preserve space before punctuation. This impacts the UHB and maybe other languages. Since the UHB is not changing that frequently, we run the risk that the user already has downloaded the latest UHB with tC 3.2.0, and then when they upgrade to 3.2.1 (for example), it will not replace the installed UHB with the bundled version because the version numbers are the same.
A few options:
1. On D43-Catalog and unfoldingWord, we increment the version number of UHB after the new tC is released - this could still be problematic
1. in the newer tCore we could add a special case for the UHB that we force update of UHB after tC upgrade
1. add processing date to the installed bibles so that we can compare to processing date of bundled resources to determine if we need to upgrade a bible
1. on tC version update we remove all the installed bibles and auto install bundled bibles (this is what we already do for tHelps). May be overkill, but safe for users, and less work for developers
1. if tC bundled resource is newer than installed resources we remove all the installed bibles and auto install bundled bibles (this is also what we already do for tHelps). May be overkill, but safe for users, and less work for developers. This is better for developers and QA since we don't have to worry about manually deleting resources before testing.
Number 5 is my recommendation. Has advantage of most safety going forward. But slows down first installs, and users will have to download again any custom bibles. 4 and 5 are probably the least effort since we already do this for tHelps.
### Features / Specifications
<!-- What are the high level features that need to be completed? -->
- [ ]
- [ ]
- [ ]
### Definition of Done
<!-- This is all the conditions, or acceptance criteria, that a
software product must satisfy are met and ready to be
accepted by a user, customer, team, or consuming system. -->
- [ ]
- [ ]
- [ ]
<!--
## QA
### Fails
- [ ]
- [ ]
-->
## Additional Context
### Mockups
|
non_defect
|
on tc startup remove bibles processed with different usfm js version update with bundled bibles story explanation user story this is a short simple description of the feature told from the perspective of the person who desires the new capability we have made a fix to usfm js to preserve space before punctuation this impacts the uhb and maybe other languages since the uhb is not changing that frequently we run the risk that the user already has downloaded the latest uhb with tc and then when they upgrade to for example it will not replace the installed uhb with the bundled version because the version numbers are the same a few options on catalog and unfoldingword we increment the version number of uhb after the new tc is released this could still be problematic in the newer tcore we could add a special case for the uhb that we force update of uhb after tc upgrade add processing date to the installed bibles so that we can compare to processing date of bundled resources to determine if we need to upgrade a bible on tc version update we remove all the installed bibles and auto install bundled bibles this is what we already do for thelps may be overkill but safe for users and less work for developers if tc bundled resource is newer than installed resources we remove all the installed bibles and auto install bundled bibles this is also what we already do for thelps may be overkill but safe for users and less work for developers this is better for developers and qa since we don t have to worry about manually deleting resources before testing number is my recommendation has advantage of most safety going forward but slows down first installs and users will have to download again any custom bibles and are probably the least effort since we already do this for thelps features specifications definition of done this is all the conditions or acceptance criteria that a software product must satisfy are met and ready to be accepted by a user customer team or consuming system qa fails additional context mockups
| 0
|
112,353
| 17,089,565,481
|
IssuesEvent
|
2021-07-08 15:41:09
|
turkdevops/atom
|
https://api.github.com/repos/turkdevops/atom
|
opened
|
CVE-2019-16776 (High) detected in npm-6.2.0.tgz
|
security vulnerability
|
## CVE-2019-16776 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-6.2.0.tgz</b></p></summary>
<p>a package manager for JavaScript</p>
<p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-6.2.0.tgz">https://registry.npmjs.org/npm/-/npm-6.2.0.tgz</a></p>
<p>
Dependency Hierarchy:
- atom-package-manager-2.4.5.tgz (Root Library)
- :x: **npm-6.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/atom/commit/b5b707a3090a251254d289cc24069b724015cea2">b5b707a3090a251254d289cc24069b724015cea2</a></p>
<p>Found in base branch: <b>electron-upgrade</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.
<p>Publish Date: 2019-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p>
<p>Release Date: 2019-12-13</p>
<p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-16776 (High) detected in npm-6.2.0.tgz - ## CVE-2019-16776 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-6.2.0.tgz</b></p></summary>
<p>a package manager for JavaScript</p>
<p>Library home page: <a href="https://registry.npmjs.org/npm/-/npm-6.2.0.tgz">https://registry.npmjs.org/npm/-/npm-6.2.0.tgz</a></p>
<p>
Dependency Hierarchy:
- atom-package-manager-2.4.5.tgz (Root Library)
- :x: **npm-6.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/atom/commit/b5b707a3090a251254d289cc24069b724015cea2">b5b707a3090a251254d289cc24069b724015cea2</a></p>
<p>Found in base branch: <b>electron-upgrade</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the npm CLI prior to 6.13.3 are vulnerable to an Arbitrary File Write. It fails to prevent access to folders outside of the intended node_modules folder through the bin field. A properly constructed entry in the package.json bin field would allow a package publisher to modify and/or gain access to arbitrary files on a user's system when the package is installed. This behavior is still possible through install scripts. This vulnerability bypasses a user using the --ignore-scripts install option.
<p>Publish Date: 2019-12-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16776>CVE-2019-16776</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli">https://blog.npmjs.org/post/189618601100/binary-planting-with-the-npm-cli</a></p>
<p>Release Date: 2019-12-13</p>
<p>Fix Resolution: npm - 6.13.3;yarn - 1.21.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in npm tgz cve high severity vulnerability vulnerable library npm tgz a package manager for javascript library home page a href dependency hierarchy atom package manager tgz root library x npm tgz vulnerable library found in head commit a href found in base branch electron upgrade vulnerability details versions of the npm cli prior to are vulnerable to an arbitrary file write it fails to prevent access to folders outside of the intended node modules folder through the bin field a properly constructed entry in the package json bin field would allow a package publisher to modify and or gain access to arbitrary files on a user s system when the package is installed this behavior is still possible through install scripts this vulnerability bypasses a user using the ignore scripts install option publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution npm yarn step up your open source security game with whitesource
| 0
|
71,148
| 23,469,479,583
|
IssuesEvent
|
2022-08-16 20:12:01
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
opened
|
Update the color of description text on Vet Center Cards
|
Needs refining ⭐️ Sitewide CMS 508/Accessibility 508-defect-2
|
## Description
On the Vet Center Dashboard in the CMS, the description text on the cards (HEX 757576) does not pass WCAG 2.0 AA color contrast requirements against the light gray background
## Screenshot

## Accessibility Standard
WCAG version 2.0 AAA, [Criterion 1.4.6](https://www.w3.org/WAI/WCAG21/Understanding/contrast-enhanced)
## Acceptance Criteria
- [ ] Interactive Design review to determine what color this should be updated to
- [ ] Technical review
- [ ] Change management is consulted
- [ ] Implementation ticket is created
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
1.0
|
Update the color of description text on Vet Center Cards - ## Description
On the Vet Center Dashboard in the CMS, the description text on the cards (HEX 757576) does not pass WCAG 2.0 AA color contrast requirements against the light gray background
## Screenshot

## Accessibility Standard
WCAG version 2.0 AAA, [Criterion 1.4.6](https://www.w3.org/WAI/WCAG21/Understanding/contrast-enhanced)
## Acceptance Criteria
- [ ] Interactive Design review to determine what color this should be updated to
- [ ] Technical review
- [ ] Change management is consulted
- [ ] Implementation ticket is created
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
|
defect
|
update the color of description text on vet center cards description on the vet center dashboard in the cms the description text on the cards hex does not pass wcag aa color contrast requirements against the light gray background screenshot accessibility standard wcag version aaa acceptance criteria interactive design review to determine what color this should be updated to technical review change management is consulted implementation ticket is created cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
| 1
|
240,034
| 19,988,374,355
|
IssuesEvent
|
2022-01-31 00:42:56
|
backend-br/vagas
|
https://api.github.com/repos/backend-br/vagas
|
closed
|
[Remoto] Python Developer - PL
|
CLT Pleno Python Django Testes automatizados PostgreSQL Git Rest Linux Docker Stale
|
## Descrição da vaga:
**Selecionamos para empresa de Tecnologia da Informação de Porto Alegre / RS**
- Inglês avançado ou fluente, falado e escrito
## Características da nossa empresa:
- Ambiente sem burocracia e descontraído;
- Poder influenciador - decidir diretamente nos resultados do trabalho;
- Trabalhar com uma equipe multinacional de entusiastas em tecnologia;
- Uso diário da língua inglesa;
- Cultura de trabalho com foco no equilíbrio entre vida pessoal e vida profissional;
## Competências gerais:
- Comunicar-se de forma eficaz e transmitir informações de forma clara para colegas e clientes;
- Participar de reuniões com clientes e entender o problema que deve ser resolvido;
- Sentir-se confortável trabalhando em um ambiente internacional;
- Autonomia, independência e automotivação para trabalhar sem gestão direta.
## Requisitos
- Pelo menos 3 anos de experiência profissional em programação em Python;
- Experiência mínima de 3 anos em desenvolvimento web;
- Sentir-se à vontade para escrever testes automatizados;
- Experiência na utilização de Git;
- Trabalhar com desenvolvimento web utilizando Django;
- Implantar e manter aplicativos no Google Cloud Platform;
- Python e Django e algum código escrito em Golang Docker / Kubernetes PostgreSQL Elastic Search
## Benefícios
- Vale-refeição R$ 28,00 x “n” dias úteis (desconto mensal simbólico de R$1,00);
- Vale-transporte (opcional, desconto de 6% em folha);
- Plano de saúde Unimed. Cobrimos 90% do valor além de 1 consulta a especialista por mês*;
- Plano de assistência odontológica Unimed. Valor mensal 90% pago pela empresa*;
- Bônus R$2.000 pela indicação de profissionais (após efetivados 90 dias);
- Bônus R$3.000 pela indicação de profissionais do sexo feminino (após efetivados 90 dias);
- Direito a férias antes de completar o 1º ano na empresa;
- Bolsa de 50% para cursos de curta duração (até 1 ano no valor máximo de R$350,00 mensais);
- Ajuda de custo de 50% para internet residencial;
- Feedback constante e avaliação de desempenho.
- Obs: Adesão Unimed e Uniodonto após o período de experiência.
## Contratação
**Regime CLT**
## Como se candidatar
**Por favor envie um email informando sua pretensão salarial para leandro.lemos@grupodreamwork.com.br com seu CV anexado - enviar no assunto: [Remoto] Python Developer - PL**
**Ou candidatar-se através do link: https://menvievagas.com.br/vaga/23517**
|
1.0
|
[Remoto] Python Developer - PL - ## Descrição da vaga:
**Selecionamos para empresa de Tecnologia da Informação de Porto Alegre / RS**
- Inglês avançado ou fluente, falado e escrito
## Características da nossa empresa:
- Ambiente sem burocracia e descontraído;
- Poder influenciador - decidir diretamente nos resultados do trabalho;
- Trabalhar com uma equipe multinacional de entusiastas em tecnologia;
- Uso diário da língua inglesa;
- Cultura de trabalho com foco no equilíbrio entre vida pessoal e vida profissional;
## Competências gerais:
- Comunicar-se de forma eficaz e transmitir informações de forma clara para colegas e clientes;
- Participar de reuniões com clientes e entender o problema que deve ser resolvido;
- Sentir-se confortável trabalhando em um ambiente internacional;
- Autonomia, independência e automotivação para trabalhar sem gestão direta.
## Requisitos
- Pelo menos 3 anos de experiência profissional em programação em Python;
- Experiência mínima de 3 anos em desenvolvimento web;
- Sentir-se à vontade para escrever testes automatizados;
- Experiência na utilização de Git;
- Trabalhar com desenvolvimento web utilizando Django;
- Implantar e manter aplicativos no Google Cloud Platform;
- Python e Django e algum código escrito em Golang Docker / Kubernetes PostgreSQL Elastic Search
## Benefícios
- Vale-refeição R$ 28,00 x “n” dias úteis (desconto mensal simbólico de R$1,00);
- Vale-transporte (opcional, desconto de 6% em folha);
- Plano de saúde Unimed. Cobrimos 90% do valor além de 1 consulta a especialista por mês*;
- Plano de assistência odontológica Unimed. Valor mensal 90% pago pela empresa*;
- Bônus R$2.000 pela indicação de profissionais (após efetivados 90 dias);
- Bônus R$3.000 pela indicação de profissionais do sexo feminino (após efetivados 90 dias);
- Direito a férias antes de completar o 1º ano na empresa;
- Bolsa de 50% para cursos de curta duração (até 1 ano no valor máximo de R$350,00 mensais);
- Ajuda de custo de 50% para internet residencial;
- Feedback constante e avaliação de desempenho.
- Obs: Adesão Unimed e Uniodonto após o período de experiência.
## Contratação
**Regime CLT**
## Como se candidatar
**Por favor envie um email informando sua pretensão salarial para leandro.lemos@grupodreamwork.com.br com seu CV anexado - enviar no assunto: [Remoto] Python Developer - PL**
**Ou candidatar-se através do link: https://menvievagas.com.br/vaga/23517**
|
non_defect
|
python developer pl descrição da vaga selecionamos para empresa de tecnologia da informação de porto alegre rs inglês avançado ou fluente falado e escrito características da nossa empresa ambiente sem burocracia e descontraído poder influenciador decidir diretamente nos resultados do trabalho trabalhar com uma equipe multinacional de entusiastas em tecnologia uso diário da língua inglesa cultura de trabalho com foco no equilíbrio entre vida pessoal e vida profissional competências gerais comunicar se de forma eficaz e transmitir informações de forma clara para colegas e clientes participar de reuniões com clientes e entender o problema que deve ser resolvido sentir se confortável trabalhando em um ambiente internacional autonomia independência e automotivação para trabalhar sem gestão direta requisitos pelo menos anos de experiência profissional em programação em python experiência mínima de anos em desenvolvimento web sentir se à vontade para escrever testes automatizados experiência na utilização de git trabalhar com desenvolvimento web utilizando django implantar e manter aplicativos no google cloud platform python e django e algum código escrito em golang docker kubernetes postgresql elastic search benefícios vale refeição r x “n” dias úteis desconto mensal simbólico de r vale transporte opcional desconto de em folha plano de saúde unimed cobrimos do valor além de consulta a especialista por mês plano de assistência odontológica unimed valor mensal pago pela empresa bônus r pela indicação de profissionais após efetivados dias bônus r pela indicação de profissionais do sexo feminino após efetivados dias direito a férias antes de completar o ano na empresa bolsa de para cursos de curta duração até ano no valor máximo de r mensais ajuda de custo de para internet residencial feedback constante e avaliação de desempenho obs adesão unimed e uniodonto após o período de experiência contratação regime clt como se candidatar por favor envie um email informando sua pretensão salarial para leandro lemos grupodreamwork com br com seu cv anexado enviar no assunto python developer pl ou candidatar se através do link
| 0
|
45,896
| 13,055,817,159
|
IssuesEvent
|
2020-07-30 02:49:32
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
NFE calling PushFrame multiple time (Trac #212)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
Migrated from https://code.icecube.wisc.edu/ticket/212
```json
{
"status": "closed",
"changetime": "2010-11-10T11:29:13",
"description": "If the NFEPulseMerger is used, along with poorly configured \nNFE settings (so no pulses generated), you get multiple calls to \nPushFrame() in the NFEPulseMerger. \n\nProblem lines are:\n\n(NFE/releases/V01-01-00)\n\nIndex: ../src/NFE/private/NFE/I3NFEPulseMerger.cxx\n===================================================================\n--- ../src/NFE/private/NFE/I3NFEPulseMerger.cxx\t(revision 68412)\n+++ ../src/NFE/private/NFE/I3NFEPulseMerger.cxx\t(working copy)\n@@ -295,7 +295,7 @@\n basePulseMapName.c_str());\n }\n ++noPulseMapCounter_;\n- PushFrame(frame, \"OutBox\");\n+ //PushFrame(frame, \"OutBox\");\n return;\n }\n \n@@ -354,7 +354,7 @@\n addPulseMapName.c_str());\n }\n ++noPulseMapCounter_;\n- PushFrame(frame, \"OutBox\");\n+ //PushFrame(frame, \"OutBox\");\n return;\n }\n\n\nThis script will trigger the bug. If the inputs to the NFE module are corrected, it runs fine.\n\n#!/usr/bin/env python \n\nfrom I3Tray import *\nimport os,glob,stat,time, sys\n\nload(\"libdataio\")\nload(\"libphys-services\")\nload(\"libpayload-parsing\")\nload(\"libdaq-decode\")\nload(\"libjebclasses\")\nload(\"libI3Db\")\nload(\"libicepick\")\n#load(\"libtwr-decode\")\nload(\"libDOMcalibrator\")\nload(\"liblinefit\")\nload(\"libNFE\")\nload(\"libDomTools\")\nload(\"liblilliput\")\nload(\"libgulliver\")\nload(\"libgulliver-modules\")\n\n\nfilename = sys.argv[1]\noutfile = sys.argv[2]\ndbserver = \"dbs2.icecube.wisc.edu\"\ndbuser = 'www'\n\ntray = I3Tray()\n\ntray.AddService(\"I3ReaderServiceFactory\",\"reader\")(\n\t(\"Filename\",filename),\n#\t(\"SkipKeys\",[\"I3DST10\",\"I3DST10Header\"]),\n\t(\"OmitGeometry\",True),\n\t(\"OmitCalibration\",True),\n\t(\"OmitStatus\",True),\n\t(\"SkipMissingDrivingTime\",True),\n\t)\n\ntray.AddService(\"I3DbOMKey2MBIDFactory\",\"omkey2mbid\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t)\n\ntray.AddService(\"I3DbGeometryServiceFactory\",\"geometry\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t(\"CompleteGeometry\",False),\n\t)\n\ntray.AddService(\"I3DbCalibrationServiceFactory\",\"calibration\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t)\n\ntray.AddService(\"I3DbDetectorStatusServiceFactory\",\"status\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t)\n\ntray.AddService(\"I3PayloadParsingEventDecoderFactory\",\"decode\")(\n (\"CPUDataID\",\"BeaconHits\"), #Cpu data ID\n (\"FlasherDataID\",\"Flasher\"), #Flasher ID.. not decoding\n (\"IceTopID\",\"IceTop\"),\n (\"InIceID\",\"InIce\"),\n (\"MinBiasID\",\"MinBias\"),\n (\"OnboardLEDID\",\"\"),\n (\"SPEDataID\",\"RawData\"),\n (\"TestDataID\",\"\"),\n (\"headerid\",\"I3DAQEventHeader\"),\n (\"triggerid\",\"I3TriggerHierarchy\"),\n (\"specialdataid\",\"I3DAQSpecialData\"),\n (\"specialdataoms\",[OMKey(0,91),OMKey(0,92)])\n )\n\ntray.AddModule(\"I3Muxer\",\"muxer\")\n\ntray.AddModule(\"I3FrameBufferDecode\",\"bufferdecode\")(\n\t(\"bufferid\",\"I3DAQData\"),\n\t)\n\ntray.AddModule(\"I3DOMLaunchCleaning\",\"4_BadDomCleaning\")( \n\t(\"InIceInput\",\"InIceRawData\"), \n\t(\"IceTopInput\",\"IceTopRawData\"), \n\t(\"InIceOutput\",\"CleanInIceRawData\"), \n\t(\"IceTopOutput\",\"CleanIceTopRawData\"), \n\t(\"FirstLaunchCleaning\",False),\n\t(\"CleanedKeys\",[OMKey(38,59),# Blackberry \n\t\t\tOMKey(6,11), # Discworld - Meteor DOM\n\t\t\tOMKey(68,42) # Krabba\n\t\t\t]) \n\t)\ntray.AddModule(\"I3LCCleaning\",\"6_InIceLCClean\")(\n\t(\"InIceInput\",\"CleanInIceRawData\"), \n\t(\"InIceOutput\",\"HLCInIceRawData\") \n\t)\n\ntray.AddModule(\"I3DOMcalibrator\",\"8_StdDomcal\")( \n\t(\"InputRawDataName\",\"HLCInIceRawData\"), \n\t(\"OutputFADCDataName\",\"CalibratedFADC\"),\n\t(\"OutputATWDDataName\",\"CalibratedATWD\"),\n\t(\"ATWDSaturationLevel\",900),\n\t(\"FADCSaturationLevel\",900),\n\t(\"CalibrateDataWithSLC\",False),\n\t(\"CalibrationMode\",0),\n\t(\"CorrectPedestalDroopDualTau\",True), \n\t(\"CorrectPedestalDroop\",False),\n\t(\"KeepCstampRedundantInfo\",False),\n\t(\"OutputToFile\",False),\n\t(\"SubtractBaseline\",False),\n\t(\"FADCTimeOffset\",0),\n\t(\"SubtractTransitTime\",True),\n\t) \n\ntray.AddService(\"I3NFEPreAlg_Eva_ServiceFactory\", \"NFEPreAlgEva\")(\n (\"ATWDSimpleThreshold\", 0.4 ), # Default\n (\"ATWDFeatureThreshold\", 0.08), # Default\n (\"ATWDFeatureMaxLength\", 6 ), # Default\n (\"ATWDFeatureMinDistance\", 4 ), # Default\n (\"FADCSimpleThreshold\", 0.6 ), # Default\n\t(\"FADCFeatureThreshold\", 0.1 ), # Default\n\t(\"FADCFeatureMaxLength\", 5 ), # Default\n\t(\"FADCFeatureMinDistance\", 3 ), # Default\n\t)\n\ntray.AddService(\"I3NFEAlg_Simple_ServiceFactory\", \"NFEAlgSimple\")(\n (\"ATWDDetectionThreshold\", 0.04 ), # Default\n (\"ATWDFeatureThreshold\", 0.04 ), # Default\n (\"ATWDDeltaTNewToroid\", -0.455 *I3Units.ns), # Default\n (\"ATWDDeltaTOldToroid\", -0.820 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP0NewToroid\", 1.54 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP1NewToroid\", 1.7 ), # Default\n (\"ATWDQTCorrelationP2NewToroid\", 0.014 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP0OldToroid\", 1.27 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP1OldToroid\", 1.7 ), # Default\n (\"ATWDQTCorrelationP2OldToroid\", 0.014 *I3Units.ns), # Default\n (\"FADCDetectionThreshold\", 0.08 ), # Default\n (\"FADCFeatureThreshold\", 0.08 ), # Default\n (\"FADCDeltaT\", 3.944 *I3Units.ns), # Default\n (\"FADCQTCorrelationP0\", 7.20 *I3Units.ns), # Default\n (\"FADCQTCorrelationP1\", 2.0 ), # Default\n (\"FADCQTCorrelationP2\", 0.005 *I3Units.ns), # Default\n (\"MinCharge\", 0.15 ), # Default\n )\n\ntray.AddService(\"I3NFEAlg_BayesUnfold_ServiceFactory\", \"NFEAlgBU\")(\n (\"MaxIterations\", 40 ), # Default\n (\"MinRelativeChange\", 0.012 ), # Default\n (\"SPEThreshold\", 0.25 ), # Default\n )\n\ntray.AddService(\"I3NFEAlg_SLCHE_ServiceFactory\", \"NFEAlgSLCHE\")(\n (\"ChargeCalibConst\", 1.23929 ), # Default\n (\"MeanParabolaArea\", 53.09 *I3Units.ns), # Default\n (\"SLCDeltaT\", -51.68 *I3Units.ns), # Default\n (\"ColinearityFactor\", 1.75 ), # Default\n )\n\n# Extract ATWD waveforms\ntray.AddModule(\"I3NFE\", \"NFEModuleATWD\")(\n (\"InputWaveformName\", \"OfflineCalibratedATWD\"), # ! Cleaned AWTD waveforms\n (\"OutputPulseMapName\", \"NFEPulsesATWD\"), # ! Output ATWD PulseMapSeries\n (\"WriteExtraInfo\", True), # ! Create the ExtraInfo map frame object\n (\"WriteAlgInfo\", False), # Default\n (\"AddEmptyPulses\", False), # Default\n (\"EnforcePulse\", True), # Default\n (\"UseDroopParameterForOTNT\", True), # Default\n (\"PreEvaluationServiceName\", \"NFEPreAlgEva\"), # Setting the default alg manually\n\t(\"SimpleAlgorithmServiceName\", \"NFEAlgSimple\"), # Setting the default alg manually\n (\"ComplexAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n (\"SLCAlgorithmServicename\", \"NFEAlgSLCHE\"), # Setting the default alg manually\n (\"EnforceAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n )\n\n# Extract FADC waveforms\ntray.AddModule(\"I3NFE\", \"NFEModuleFADC\")(\n (\"InputWaveformName\", \"OfflineCalibratedFADC\"), # ! Cleaned FADC waveforms\n (\"OutputPulseMapName\", \"NFEPulsesFADC\"), # ! Output FADC PulseMapSeries\n (\"WriteExtraInfo\", True), # ! Create the ExtraInfo map frame object\n (\"WriteAlgInfo\", False), # Default\n (\"AddEmptyPulses\", False), # Default\n (\"EnforcePulse\", True), # Default\n (\"UseDroopParameterForOTNT\", True), # Default\n (\"PreEvaluationServiceName\", \"NFEPreAlgEva\"), # Setting the default alg manually\n (\"SimpleAlgorithmServiceName\", \"NFEAlgSimple\"), # Setting the default alg manually\n (\"ComplexAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n (\"SLCAlgorithmServicename\", \"NFEAlgSLCHE\"), # Setting the default alg manually\n (\"EnforceAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n )\n\ntray.AddModule(\"I3NFEPulseMerger\", \"NFEPulseMergerATWDFADC\")(\n\t(\"ATWDPulseMapName\", \"NFEPulsesATWD\"), # ! Input ATWD PulseMapSeries\n\t(\"FADCPulseMapName\", \"NFEPulsesFADC\"), # ! Input FADC+SLC PulseMapSeries\n\t(\"SLCPulseMapName\", \"\"), # Default, no third PulseMap needed\n\t(\"MergedPulseMapName\", \"NFEMergedPulsesHLC\"), # ! Output ATWD+FADC PulseMapSeries\n\t(\"AlreadySorted\", True), # ! minor performance optimization\n\t(\"WriteExtraInfo\", True), # ! Create the ExtraInfo map frame object\n\t(\"WriteAlgInfo\", False), # Default\n\t(\"ExclusionTime\", 15. *I3Units.ns), # Default\n\t)\ntray.AddModule(\"I3TimeWindowCleaning<I3RecoPulse>\",\"timewindwoPulses\")(\n \t(\"InputResponse\",\"NFEMergedPulsesHLC\"),\n \t(\"OutputResponse\",\"TWCMuonPulseSeriesReco\"),\n \t(\"TimeWindow\",6000)\n \t)\ntray.AddModule(\"Dump\",'dump')\ntray.AddModule(\"TrashCan\",\"can\")\n\ntray.Execute()\ntray.Finish()\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1289388553000000",
"component": "combo reconstruction",
"summary": "NFE calling PushFrame multiple time",
"priority": "normal",
"keywords": "",
"time": "2010-11-09T18:31:29",
"milestone": "",
"owner": "mwallraff",
"type": "defect"
}
```
|
1.0
|
NFE calling PushFrame multiple time (Trac #212) - Migrated from https://code.icecube.wisc.edu/ticket/212
```json
{
"status": "closed",
"changetime": "2010-11-10T11:29:13",
"description": "If the NFEPulseMerger is used, along with poorly configured \nNFE settings (so no pulses generated), you get multiple calls to \nPushFrame() in the NFEPulseMerger. \n\nProblem lines are:\n\n(NFE/releases/V01-01-00)\n\nIndex: ../src/NFE/private/NFE/I3NFEPulseMerger.cxx\n===================================================================\n--- ../src/NFE/private/NFE/I3NFEPulseMerger.cxx\t(revision 68412)\n+++ ../src/NFE/private/NFE/I3NFEPulseMerger.cxx\t(working copy)\n@@ -295,7 +295,7 @@\n basePulseMapName.c_str());\n }\n ++noPulseMapCounter_;\n- PushFrame(frame, \"OutBox\");\n+ //PushFrame(frame, \"OutBox\");\n return;\n }\n \n@@ -354,7 +354,7 @@\n addPulseMapName.c_str());\n }\n ++noPulseMapCounter_;\n- PushFrame(frame, \"OutBox\");\n+ //PushFrame(frame, \"OutBox\");\n return;\n }\n\n\nThis script will trigger the bug. If the inputs to the NFE module are corrected, it runs fine.\n\n#!/usr/bin/env python \n\nfrom I3Tray import *\nimport os,glob,stat,time, sys\n\nload(\"libdataio\")\nload(\"libphys-services\")\nload(\"libpayload-parsing\")\nload(\"libdaq-decode\")\nload(\"libjebclasses\")\nload(\"libI3Db\")\nload(\"libicepick\")\n#load(\"libtwr-decode\")\nload(\"libDOMcalibrator\")\nload(\"liblinefit\")\nload(\"libNFE\")\nload(\"libDomTools\")\nload(\"liblilliput\")\nload(\"libgulliver\")\nload(\"libgulliver-modules\")\n\n\nfilename = sys.argv[1]\noutfile = sys.argv[2]\ndbserver = \"dbs2.icecube.wisc.edu\"\ndbuser = 'www'\n\ntray = I3Tray()\n\ntray.AddService(\"I3ReaderServiceFactory\",\"reader\")(\n\t(\"Filename\",filename),\n#\t(\"SkipKeys\",[\"I3DST10\",\"I3DST10Header\"]),\n\t(\"OmitGeometry\",True),\n\t(\"OmitCalibration\",True),\n\t(\"OmitStatus\",True),\n\t(\"SkipMissingDrivingTime\",True),\n\t)\n\ntray.AddService(\"I3DbOMKey2MBIDFactory\",\"omkey2mbid\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t)\n\ntray.AddService(\"I3DbGeometryServiceFactory\",\"geometry\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t(\"CompleteGeometry\",False),\n\t)\n\ntray.AddService(\"I3DbCalibrationServiceFactory\",\"calibration\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t)\n\ntray.AddService(\"I3DbDetectorStatusServiceFactory\",\"status\")(\n\t(\"host\",dbserver),\n\t(\"username\",dbuser),\n\t(\"database\",\"I3OmDb\"),\n\t)\n\ntray.AddService(\"I3PayloadParsingEventDecoderFactory\",\"decode\")(\n (\"CPUDataID\",\"BeaconHits\"), #Cpu data ID\n (\"FlasherDataID\",\"Flasher\"), #Flasher ID.. not decoding\n (\"IceTopID\",\"IceTop\"),\n (\"InIceID\",\"InIce\"),\n (\"MinBiasID\",\"MinBias\"),\n (\"OnboardLEDID\",\"\"),\n (\"SPEDataID\",\"RawData\"),\n (\"TestDataID\",\"\"),\n (\"headerid\",\"I3DAQEventHeader\"),\n (\"triggerid\",\"I3TriggerHierarchy\"),\n (\"specialdataid\",\"I3DAQSpecialData\"),\n (\"specialdataoms\",[OMKey(0,91),OMKey(0,92)])\n )\n\ntray.AddModule(\"I3Muxer\",\"muxer\")\n\ntray.AddModule(\"I3FrameBufferDecode\",\"bufferdecode\")(\n\t(\"bufferid\",\"I3DAQData\"),\n\t)\n\ntray.AddModule(\"I3DOMLaunchCleaning\",\"4_BadDomCleaning\")( \n\t(\"InIceInput\",\"InIceRawData\"), \n\t(\"IceTopInput\",\"IceTopRawData\"), \n\t(\"InIceOutput\",\"CleanInIceRawData\"), \n\t(\"IceTopOutput\",\"CleanIceTopRawData\"), \n\t(\"FirstLaunchCleaning\",False),\n\t(\"CleanedKeys\",[OMKey(38,59),# Blackberry \n\t\t\tOMKey(6,11), # Discworld - Meteor DOM\n\t\t\tOMKey(68,42) # Krabba\n\t\t\t]) \n\t)\ntray.AddModule(\"I3LCCleaning\",\"6_InIceLCClean\")(\n\t(\"InIceInput\",\"CleanInIceRawData\"), \n\t(\"InIceOutput\",\"HLCInIceRawData\") \n\t)\n\ntray.AddModule(\"I3DOMcalibrator\",\"8_StdDomcal\")( \n\t(\"InputRawDataName\",\"HLCInIceRawData\"), \n\t(\"OutputFADCDataName\",\"CalibratedFADC\"),\n\t(\"OutputATWDDataName\",\"CalibratedATWD\"),\n\t(\"ATWDSaturationLevel\",900),\n\t(\"FADCSaturationLevel\",900),\n\t(\"CalibrateDataWithSLC\",False),\n\t(\"CalibrationMode\",0),\n\t(\"CorrectPedestalDroopDualTau\",True), \n\t(\"CorrectPedestalDroop\",False),\n\t(\"KeepCstampRedundantInfo\",False),\n\t(\"OutputToFile\",False),\n\t(\"SubtractBaseline\",False),\n\t(\"FADCTimeOffset\",0),\n\t(\"SubtractTransitTime\",True),\n\t) \n\ntray.AddService(\"I3NFEPreAlg_Eva_ServiceFactory\", \"NFEPreAlgEva\")(\n (\"ATWDSimpleThreshold\", 0.4 ), # Default\n (\"ATWDFeatureThreshold\", 0.08), # Default\n (\"ATWDFeatureMaxLength\", 6 ), # Default\n (\"ATWDFeatureMinDistance\", 4 ), # Default\n (\"FADCSimpleThreshold\", 0.6 ), # Default\n\t(\"FADCFeatureThreshold\", 0.1 ), # Default\n\t(\"FADCFeatureMaxLength\", 5 ), # Default\n\t(\"FADCFeatureMinDistance\", 3 ), # Default\n\t)\n\ntray.AddService(\"I3NFEAlg_Simple_ServiceFactory\", \"NFEAlgSimple\")(\n (\"ATWDDetectionThreshold\", 0.04 ), # Default\n (\"ATWDFeatureThreshold\", 0.04 ), # Default\n (\"ATWDDeltaTNewToroid\", -0.455 *I3Units.ns), # Default\n (\"ATWDDeltaTOldToroid\", -0.820 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP0NewToroid\", 1.54 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP1NewToroid\", 1.7 ), # Default\n (\"ATWDQTCorrelationP2NewToroid\", 0.014 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP0OldToroid\", 1.27 *I3Units.ns), # Default\n (\"ATWDQTCorrelationP1OldToroid\", 1.7 ), # Default\n (\"ATWDQTCorrelationP2OldToroid\", 0.014 *I3Units.ns), # Default\n (\"FADCDetectionThreshold\", 0.08 ), # Default\n (\"FADCFeatureThreshold\", 0.08 ), # Default\n (\"FADCDeltaT\", 3.944 *I3Units.ns), # Default\n (\"FADCQTCorrelationP0\", 7.20 *I3Units.ns), # Default\n (\"FADCQTCorrelationP1\", 2.0 ), # Default\n (\"FADCQTCorrelationP2\", 0.005 *I3Units.ns), # Default\n (\"MinCharge\", 0.15 ), # Default\n )\n\ntray.AddService(\"I3NFEAlg_BayesUnfold_ServiceFactory\", \"NFEAlgBU\")(\n (\"MaxIterations\", 40 ), # Default\n (\"MinRelativeChange\", 0.012 ), # Default\n (\"SPEThreshold\", 0.25 ), # Default\n )\n\ntray.AddService(\"I3NFEAlg_SLCHE_ServiceFactory\", \"NFEAlgSLCHE\")(\n (\"ChargeCalibConst\", 1.23929 ), # Default\n (\"MeanParabolaArea\", 53.09 *I3Units.ns), # Default\n (\"SLCDeltaT\", -51.68 *I3Units.ns), # Default\n (\"ColinearityFactor\", 1.75 ), # Default\n )\n\n# Extract ATWD waveforms\ntray.AddModule(\"I3NFE\", \"NFEModuleATWD\")(\n (\"InputWaveformName\", \"OfflineCalibratedATWD\"), # ! Cleaned AWTD waveforms\n (\"OutputPulseMapName\", \"NFEPulsesATWD\"), # ! Output ATWD PulseMapSeries\n (\"WriteExtraInfo\", True), # ! Create the ExtraInfo map frame object\n (\"WriteAlgInfo\", False), # Default\n (\"AddEmptyPulses\", False), # Default\n (\"EnforcePulse\", True), # Default\n (\"UseDroopParameterForOTNT\", True), # Default\n (\"PreEvaluationServiceName\", \"NFEPreAlgEva\"), # Setting the default alg manually\n\t(\"SimpleAlgorithmServiceName\", \"NFEAlgSimple\"), # Setting the default alg manually\n (\"ComplexAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n (\"SLCAlgorithmServicename\", \"NFEAlgSLCHE\"), # Setting the default alg manually\n (\"EnforceAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n )\n\n# Extract FADC waveforms\ntray.AddModule(\"I3NFE\", \"NFEModuleFADC\")(\n (\"InputWaveformName\", \"OfflineCalibratedFADC\"), # ! Cleaned FADC waveforms\n (\"OutputPulseMapName\", \"NFEPulsesFADC\"), # ! Output FADC PulseMapSeries\n (\"WriteExtraInfo\", True), # ! Create the ExtraInfo map frame object\n (\"WriteAlgInfo\", False), # Default\n (\"AddEmptyPulses\", False), # Default\n (\"EnforcePulse\", True), # Default\n (\"UseDroopParameterForOTNT\", True), # Default\n (\"PreEvaluationServiceName\", \"NFEPreAlgEva\"), # Setting the default alg manually\n (\"SimpleAlgorithmServiceName\", \"NFEAlgSimple\"), # Setting the default alg manually\n (\"ComplexAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n (\"SLCAlgorithmServicename\", \"NFEAlgSLCHE\"), # Setting the default alg manually\n (\"EnforceAlgorithmServiceName\", \"NFEAlgBU\"), # Setting the default alg manually\n )\n\ntray.AddModule(\"I3NFEPulseMerger\", \"NFEPulseMergerATWDFADC\")(\n\t(\"ATWDPulseMapName\", \"NFEPulsesATWD\"), # ! Input ATWD PulseMapSeries\n\t(\"FADCPulseMapName\", \"NFEPulsesFADC\"), # ! Input FADC+SLC PulseMapSeries\n\t(\"SLCPulseMapName\", \"\"), # Default, no third PulseMap needed\n\t(\"MergedPulseMapName\", \"NFEMergedPulsesHLC\"), # ! Output ATWD+FADC PulseMapSeries\n\t(\"AlreadySorted\", True), # ! minor performance optimization\n\t(\"WriteExtraInfo\", True), # ! Create the ExtraInfo map frame object\n\t(\"WriteAlgInfo\", False), # Default\n\t(\"ExclusionTime\", 15. *I3Units.ns), # Default\n\t)\ntray.AddModule(\"I3TimeWindowCleaning<I3RecoPulse>\",\"timewindwoPulses\")(\n \t(\"InputResponse\",\"NFEMergedPulsesHLC\"),\n \t(\"OutputResponse\",\"TWCMuonPulseSeriesReco\"),\n \t(\"TimeWindow\",6000)\n \t)\ntray.AddModule(\"Dump\",'dump')\ntray.AddModule(\"TrashCan\",\"can\")\n\ntray.Execute()\ntray.Finish()\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1289388553000000",
"component": "combo reconstruction",
"summary": "NFE calling PushFrame multiple time",
"priority": "normal",
"keywords": "",
"time": "2010-11-09T18:31:29",
"milestone": "",
"owner": "mwallraff",
"type": "defect"
}
```
|
defect
|
nfe calling pushframe multiple time trac migrated from json status closed changetime description if the nfepulsemerger is used along with poorly configured nnfe settings so no pulses generated you get multiple calls to npushframe in the nfepulsemerger n nproblem lines are n n nfe releases n nindex src nfe private nfe cxx n n src nfe private nfe cxx t revision n src nfe private nfe cxx t working copy n n basepulsemapname c str n n nopulsemapcounter n pushframe frame outbox n pushframe frame outbox n return n n n n addpulsemapname c str n n nopulsemapcounter n pushframe frame outbox n pushframe frame outbox n return n n n nthis script will trigger the bug if the inputs to the nfe module are corrected it runs fine n n usr bin env python n nfrom import nimport os glob stat time sys n nload libdataio nload libphys services nload libpayload parsing nload libdaq decode nload libjebclasses nload nload libicepick n load libtwr decode nload libdomcalibrator nload liblinefit nload libnfe nload libdomtools nload liblilliput nload libgulliver nload libgulliver modules n n nfilename sys argv noutfile sys argv ndbserver icecube wisc edu ndbuser www n ntray n ntray addservice reader n t filename filename n t skipkeys n t omitgeometry true n t omitcalibration true n t omitstatus true n t skipmissingdrivingtime true n t n ntray addservice n t host dbserver n t username dbuser n t database n t n ntray addservice geometry n t host dbserver n t username dbuser n t database n t completegeometry false n t n ntray addservice calibration n t host dbserver n t username dbuser n t database n t n ntray addservice status n t host dbserver n t username dbuser n t database n t n ntray addservice decode n cpudataid beaconhits cpu data id n flasherdataid flasher flasher id not decoding n icetopid icetop n iniceid inice n minbiasid minbias n onboardledid n spedataid rawdata n testdataid n headerid n triggerid n specialdataid n specialdataoms n n ntray addmodule muxer n ntray addmodule bufferdecode n t bufferid n t n ntray addmodule baddomcleaning n t iniceinput inicerawdata n t icetopinput icetoprawdata n t iniceoutput cleaninicerawdata n t icetopoutput cleanicetoprawdata n t firstlaunchcleaning false n t cleanedkeys n t ntray addmodule inicelcclean n t iniceinput cleaninicerawdata n t iniceoutput hlcinicerawdata n t n ntray addmodule stddomcal n t inputrawdataname hlcinicerawdata n t outputfadcdataname calibratedfadc n t outputatwddataname calibratedatwd n t atwdsaturationlevel n t fadcsaturationlevel n t calibratedatawithslc false n t calibrationmode n t correctpedestaldroopdualtau true n t correctpedestaldroop false n t keepcstampredundantinfo false n t outputtofile false n t subtractbaseline false n t fadctimeoffset n t subtracttransittime true n t n ntray addservice eva servicefactory nfeprealgeva n atwdsimplethreshold default n atwdfeaturethreshold default n atwdfeaturemaxlength default n atwdfeaturemindistance default n fadcsimplethreshold default n t fadcfeaturethreshold default n t fadcfeaturemaxlength default n t fadcfeaturemindistance default n t n ntray addservice simple servicefactory nfealgsimple n atwddetectionthreshold default n atwdfeaturethreshold default n atwddeltatnewtoroid ns default n atwddeltatoldtoroid ns default n ns default n default n ns default n ns default n default n ns default n fadcdetectionthreshold default n fadcfeaturethreshold default n fadcdeltat ns default n ns default n default n ns default n mincharge default n n ntray addservice bayesunfold servicefactory nfealgbu n maxiterations default n minrelativechange default n spethreshold default n n ntray addservice slche servicefactory nfealgslche n chargecalibconst default n meanparabolaarea ns default n slcdeltat ns default n colinearityfactor default n n n extract atwd waveforms ntray addmodule nfemoduleatwd n inputwaveformname offlinecalibratedatwd cleaned awtd waveforms n outputpulsemapname nfepulsesatwd output atwd pulsemapseries n writeextrainfo true create the extrainfo map frame object n writealginfo false default n addemptypulses false default n enforcepulse true default n usedroopparameterforotnt true default n preevaluationservicename nfeprealgeva setting the default alg manually n t simplealgorithmservicename nfealgsimple setting the default alg manually n complexalgorithmservicename nfealgbu setting the default alg manually n slcalgorithmservicename nfealgslche setting the default alg manually n enforcealgorithmservicename nfealgbu setting the default alg manually n n n extract fadc waveforms ntray addmodule nfemodulefadc n inputwaveformname offlinecalibratedfadc cleaned fadc waveforms n outputpulsemapname nfepulsesfadc output fadc pulsemapseries n writeextrainfo true create the extrainfo map frame object n writealginfo false default n addemptypulses false default n enforcepulse true default n usedroopparameterforotnt true default n preevaluationservicename nfeprealgeva setting the default alg manually n simplealgorithmservicename nfealgsimple setting the default alg manually n complexalgorithmservicename nfealgbu setting the default alg manually n slcalgorithmservicename nfealgslche setting the default alg manually n enforcealgorithmservicename nfealgbu setting the default alg manually n n ntray addmodule nfepulsemergeratwdfadc n t atwdpulsemapname nfepulsesatwd input atwd pulsemapseries n t fadcpulsemapname nfepulsesfadc input fadc slc pulsemapseries n t slcpulsemapname default no third pulsemap needed n t mergedpulsemapname nfemergedpulseshlc output atwd fadc pulsemapseries n t alreadysorted true minor performance optimization n t writeextrainfo true create the extrainfo map frame object n t writealginfo false default n t exclusiontime ns default n t ntray addmodule timewindwopulses n t inputresponse nfemergedpulseshlc n t outputresponse twcmuonpulseseriesreco n t timewindow n t ntray addmodule dump dump ntray addmodule trashcan can n ntray execute ntray finish n reporter blaufuss cc resolution fixed ts component combo reconstruction summary nfe calling pushframe multiple time priority normal keywords time milestone owner mwallraff type defect
| 1
|
248
| 2,521,794,947
|
IssuesEvent
|
2015-01-19 16:56:55
|
UNH-CORE/turbine-test-bed
|
https://api.github.com/repos/UNH-CORE/turbine-test-bed
|
opened
|
Trim cylinder turbulence generator 80/20 member
|
defect
|
It's a bit tight fitting inside the carriage extrusions. Trimming 1/16 to 1/8 inch should be enough.
|
1.0
|
Trim cylinder turbulence generator 80/20 member - It's a bit tight fitting inside the carriage extrusions. Trimming 1/16 to 1/8 inch should be enough.
|
defect
|
trim cylinder turbulence generator member it s a bit tight fitting inside the carriage extrusions trimming to inch should be enough
| 1
|
12,361
| 2,694,221,025
|
IssuesEvent
|
2015-04-01 19:01:14
|
JesusFreke/smali
|
https://api.github.com/repos/JesusFreke/smali
|
closed
|
Exception occured when I used "-r" option.
|
Priority-Medium Type-Defect
|
Original [issue 59](https://code.google.com/p/smali/issues/detail?id=59) created by JesusFreke on 2011-03-07T02:57:05.000Z:
<b>What seems to be the problem?</b>
I use baksmali to disassembly a dex file.
When i use the "-r ALL" option, baksmail throws a Exception.
when i erase the "-r ALL" option, baksmali works well.
whether i use the option incorrectly?
<b>What is the exact smali/baksmali command that you ran?</b>
java -jar baksmali-1.2.6.jar c:\eclipse_workspace\AndroidTest\bin\classes.dex -r ALL
<b>What version of smali/baksmali are you using? What rom are you working</b>
<b>from?</b>
baksmali-1.2.6 on winXP Professional sp3
<b>What is the airspeed velocity of an unladen swallow?</b>
<b>Please provide any additional information below: error messages, symptoms,</b>
<b>etc.</b>
Error occured while loading boot class path files. Aborting.
org.jf.dexlib.Code.Analysis.ClassPath$ClassNotFoundException: Could not find superclass Ljunit/framework/TestCase;
at org.jf.dexlib.Code.Analysis.ClassPath$ClassDef.loadSuperclass(ClassPath.java:784)
at org.jf.dexlib.Code.Analysis.ClassPath$ClassDef.<init>(ClassPath.java:668)
at org.jf.dexlib.Code.Analysis.ClassPath.loadClassDef(ClassPath.java:280)
at org.jf.dexlib.Code.Analysis.ClassPath.initClassPath(ClassPath.java:163)
at org.jf.dexlib.Code.Analysis.ClassPath.InitializeClassPath(ClassPath.java:131)
at org.jf.baksmali.baksmali.disassembleDexFile(baksmali.java:105)
at org.jf.baksmali.main.main(main.java:278)
Error while loading class Landroid/test/AndroidTestCase; from file .\framework.zip
Error while loading ClassPath class Landroid/test/AndroidTestCase;
|
1.0
|
Exception occured when I used "-r" option. - Original [issue 59](https://code.google.com/p/smali/issues/detail?id=59) created by JesusFreke on 2011-03-07T02:57:05.000Z:
<b>What seems to be the problem?</b>
I use baksmali to disassembly a dex file.
When i use the "-r ALL" option, baksmail throws a Exception.
when i erase the "-r ALL" option, baksmali works well.
whether i use the option incorrectly?
<b>What is the exact smali/baksmali command that you ran?</b>
java -jar baksmali-1.2.6.jar c:\eclipse_workspace\AndroidTest\bin\classes.dex -r ALL
<b>What version of smali/baksmali are you using? What rom are you working</b>
<b>from?</b>
baksmali-1.2.6 on winXP Professional sp3
<b>What is the airspeed velocity of an unladen swallow?</b>
<b>Please provide any additional information below: error messages, symptoms,</b>
<b>etc.</b>
Error occured while loading boot class path files. Aborting.
org.jf.dexlib.Code.Analysis.ClassPath$ClassNotFoundException: Could not find superclass Ljunit/framework/TestCase;
at org.jf.dexlib.Code.Analysis.ClassPath$ClassDef.loadSuperclass(ClassPath.java:784)
at org.jf.dexlib.Code.Analysis.ClassPath$ClassDef.<init>(ClassPath.java:668)
at org.jf.dexlib.Code.Analysis.ClassPath.loadClassDef(ClassPath.java:280)
at org.jf.dexlib.Code.Analysis.ClassPath.initClassPath(ClassPath.java:163)
at org.jf.dexlib.Code.Analysis.ClassPath.InitializeClassPath(ClassPath.java:131)
at org.jf.baksmali.baksmali.disassembleDexFile(baksmali.java:105)
at org.jf.baksmali.main.main(main.java:278)
Error while loading class Landroid/test/AndroidTestCase; from file .\framework.zip
Error while loading ClassPath class Landroid/test/AndroidTestCase;
|
defect
|
exception occured when i used r option original created by jesusfreke on what seems to be the problem i use baksmali to disassembly a dex file when i use the quot r all quot option baksmail throws a exception when i erase the quot r all quot option baksmali works well whether i use the option incorrectly what is the exact smali baksmali command that you ran java jar baksmali jar c eclipse workspace androidtest bin classes dex r all what version of smali baksmali are you using what rom are you working from baksmali on winxp professional what is the airspeed velocity of an unladen swallow please provide any additional information below error messages symptoms etc error occured while loading boot class path files aborting org jf dexlib code analysis classpath classnotfoundexception could not find superclass ljunit framework testcase at org jf dexlib code analysis classpath classdef loadsuperclass classpath java at org jf dexlib code analysis classpath classdef lt init gt classpath java at org jf dexlib code analysis classpath loadclassdef classpath java at org jf dexlib code analysis classpath initclasspath classpath java at org jf dexlib code analysis classpath initializeclasspath classpath java at org jf baksmali baksmali disassembledexfile baksmali java at org jf baksmali main main main java error while loading class landroid test androidtestcase from file framework zip error while loading classpath class landroid test androidtestcase
| 1
|
31,275
| 5,922,031,805
|
IssuesEvent
|
2017-05-23 01:36:37
|
zulip/zulip
|
https://api.github.com/repos/zulip/zulip
|
closed
|
Update user documentation images with updated UI screenshots
|
area: documentation (user) bite size in progress
|
Various images used in the user documentation (located in `static/images/help`) are outdated due to the recent changes in the Zulip user interface. They should be updated to reflect the current UI.
|
1.0
|
Update user documentation images with updated UI screenshots - Various images used in the user documentation (located in `static/images/help`) are outdated due to the recent changes in the Zulip user interface. They should be updated to reflect the current UI.
|
non_defect
|
update user documentation images with updated ui screenshots various images used in the user documentation located in static images help are outdated due to the recent changes in the zulip user interface they should be updated to reflect the current ui
| 0
|
29,119
| 2,713,670,750
|
IssuesEvent
|
2015-04-09 20:40:37
|
nickpaventi/culligan-diy
|
https://api.github.com/repos/nickpaventi/culligan-diy
|
closed
|
Home: Learn More CTA should line up with quote
|
Low Priority
|
Left edges should be aligned. See screen shot

|
1.0
|
Home: Learn More CTA should line up with quote - Left edges should be aligned. See screen shot

|
non_defect
|
home learn more cta should line up with quote left edges should be aligned see screen shot
| 0
|
45,012
| 12,519,732,240
|
IssuesEvent
|
2020-06-03 14:51:21
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
Coil:Cooling:DX:VariableSpeed air-to-air coil autosizes cooling capacity using CapFT curve results based on RatedInletWaterTemp = 84.2
|
Defect
|
Issue overview
--------------
line 3357 of VariableSpeedCoil:
TotCapTempModFac = CurveValue( VarSpeedCoil( DXCoilNum ).MSCCapFTemp( VarSpeedCoil( DXCoilNum ).NormSpedLevel ), MixWetBulb, RatedInletWaterTemp );
Found during development of unit test = UnitarySystem_MultispeedPerformance
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
1.0
|
Coil:Cooling:DX:VariableSpeed air-to-air coil autosizes cooling capacity using CapFT curve results based on RatedInletWaterTemp = 84.2 - Issue overview
--------------
line 3357 of VariableSpeedCoil:
TotCapTempModFac = CurveValue( VarSpeedCoil( DXCoilNum ).MSCCapFTemp( VarSpeedCoil( DXCoilNum ).NormSpedLevel ), MixWetBulb, RatedInletWaterTemp );
Found during development of unit test = UnitarySystem_MultispeedPerformance
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
defect
|
coil cooling dx variablespeed air to air coil autosizes cooling capacity using capft curve results based on ratedinletwatertemp issue overview line of variablespeedcoil totcaptempmodfac curvevalue varspeedcoil dxcoilnum msccapftemp varspeedcoil dxcoilnum normspedlevel mixwetbulb ratedinletwatertemp found during development of unit test unitarysystem multispeedperformance details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
| 1
|
767,424
| 26,924,423,857
|
IssuesEvent
|
2023-02-07 12:50:19
|
canonical/canonical.com
|
https://api.github.com/repos/canonical/canonical.com
|
closed
|
The circle of products should lead to the product pages or to the description below
|
Priority: Medium
|
The first glance to the circle of elements on the right of the main page naturally makes you think that each icon should be active and leads you somewhere, or at least have some kind of a pop-up with description.
|
1.0
|
The circle of products should lead to the product pages or to the description below - The first glance to the circle of elements on the right of the main page naturally makes you think that each icon should be active and leads you somewhere, or at least have some kind of a pop-up with description.
|
non_defect
|
the circle of products should lead to the product pages or to the description below the first glance to the circle of elements on the right of the main page naturally makes you think that each icon should be active and leads you somewhere or at least have some kind of a pop up with description
| 0
|
264,940
| 8,328,513,743
|
IssuesEvent
|
2018-09-27 01:16:28
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
business.apple.com - Firefox not supported
|
browser-firefox priority-critical severity-important
|
<!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: web -->
**URL**: https://business.apple.com/
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Site does not allow firefox to use it
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2018/9/01f9c952-0f10-4adf-8874-c898bbbf9eb5.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
business.apple.com - Firefox not supported - <!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: web -->
**URL**: https://business.apple.com/
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: Site does not allow firefox to use it
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2018/9/01f9c952-0f10-4adf-8874-c898bbbf9eb5.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
business apple com firefox not supported url browser version firefox operating system windows tested another browser yes problem type site is not usable description site does not allow firefox to use it steps to reproduce from with ❤️
| 0
|
520,334
| 15,084,596,559
|
IssuesEvent
|
2021-02-05 17:23:23
|
alerj/aloalerjsite
|
https://api.github.com/repos/alerj/aloalerjsite
|
closed
|
Adicionar campo de comissão que criou um andamento
|
priority 1
|
O campo `created_by_committee_id` deve conter o departamento do atendente no momento em que gerou o andamento.
- [x] Adicionar o campo `created_by_committee_id` à tablela progresses
- [x] Fazer uma migration jogando todos os protocolos anteriores a 01/06/2020 para o ALÔ ALERJ. A inclusão dos usuários da comissão do adolescente e do idoso foi feita no dia 03/06/2020 15:13(solicitação no e-mail). Os protocolos criados depois disso devem ser preenchidos conforme a permissão do usuário que a criou. Caso o usuário tenha permissão para 'alo-alerj', incluindo a função Operador, associar ao ALÔ ALERJ. Caso seja um usuário estritamente de uma comissão, associar à comissão correspondente.
- [x] Ao criar um andamento, salvar a comissão em que ele faz parte conforme a regra acima.
- [x] Mostrar `Departamento de Origem` nas tabelas de exibição dos protocolos
- [x] Mostrar `Departamento de Origem` desabilitado no form dos protocolos
- [x] Adicionar à busca avançada um filtro `Departamento de Origem`
|
1.0
|
Adicionar campo de comissão que criou um andamento - O campo `created_by_committee_id` deve conter o departamento do atendente no momento em que gerou o andamento.
- [x] Adicionar o campo `created_by_committee_id` à tablela progresses
- [x] Fazer uma migration jogando todos os protocolos anteriores a 01/06/2020 para o ALÔ ALERJ. A inclusão dos usuários da comissão do adolescente e do idoso foi feita no dia 03/06/2020 15:13(solicitação no e-mail). Os protocolos criados depois disso devem ser preenchidos conforme a permissão do usuário que a criou. Caso o usuário tenha permissão para 'alo-alerj', incluindo a função Operador, associar ao ALÔ ALERJ. Caso seja um usuário estritamente de uma comissão, associar à comissão correspondente.
- [x] Ao criar um andamento, salvar a comissão em que ele faz parte conforme a regra acima.
- [x] Mostrar `Departamento de Origem` nas tabelas de exibição dos protocolos
- [x] Mostrar `Departamento de Origem` desabilitado no form dos protocolos
- [x] Adicionar à busca avançada um filtro `Departamento de Origem`
|
non_defect
|
adicionar campo de comissão que criou um andamento o campo created by committee id deve conter o departamento do atendente no momento em que gerou o andamento adicionar o campo created by committee id à tablela progresses fazer uma migration jogando todos os protocolos anteriores a para o alô alerj a inclusão dos usuários da comissão do adolescente e do idoso foi feita no dia solicitação no e mail os protocolos criados depois disso devem ser preenchidos conforme a permissão do usuário que a criou caso o usuário tenha permissão para alo alerj incluindo a função operador associar ao alô alerj caso seja um usuário estritamente de uma comissão associar à comissão correspondente ao criar um andamento salvar a comissão em que ele faz parte conforme a regra acima mostrar departamento de origem nas tabelas de exibição dos protocolos mostrar departamento de origem desabilitado no form dos protocolos adicionar à busca avançada um filtro departamento de origem
| 0
|
76,631
| 9,958,830,240
|
IssuesEvent
|
2019-07-05 23:44:27
|
clr1107/Vessel
|
https://api.github.com/repos/clr1107/Vessel
|
closed
|
Complete documentation
|
documentation
|
Before the new release, finish documentation and make it intuitive but short.
|
1.0
|
Complete documentation - Before the new release, finish documentation and make it intuitive but short.
|
non_defect
|
complete documentation before the new release finish documentation and make it intuitive but short
| 0
|
98,102
| 4,017,575,772
|
IssuesEvent
|
2016-05-16 05:25:36
|
xcat2/xcat-core
|
https://api.github.com/repos/xcat2/xcat-core
|
closed
|
[FVT]sles11.4 diskfull install failed on p8 ppc64
|
priority:high status:pending type:bug
|
env:sles11.4 , p8 ppc64
build:
linux-z38a:~ # lsxcatd -v
Version 2.12 (git commit 9d4c8b9c09bf0afabad32110ab27ade6d4ca9c28, built Thu May 5 07:58:26 EDT 2016)
How to reproduce:
```
During diskfull provision:
[](url)
Loading Install kernel ...
Loading initial ramdisk ...
SuSE Linux zImage starting: loaded at 086d0000-0a7e7d30 (0/0/00c3da70; sp: 06a3fe10)
uncompressing ELF header done. (00000100 bytes)
Can't allocate memory for kernel image !
EXIT called ok
```
|
1.0
|
[FVT]sles11.4 diskfull install failed on p8 ppc64 - env:sles11.4 , p8 ppc64
build:
linux-z38a:~ # lsxcatd -v
Version 2.12 (git commit 9d4c8b9c09bf0afabad32110ab27ade6d4ca9c28, built Thu May 5 07:58:26 EDT 2016)
How to reproduce:
```
During diskfull provision:
[](url)
Loading Install kernel ...
Loading initial ramdisk ...
SuSE Linux zImage starting: loaded at 086d0000-0a7e7d30 (0/0/00c3da70; sp: 06a3fe10)
uncompressing ELF header done. (00000100 bytes)
Can't allocate memory for kernel image !
EXIT called ok
```
|
non_defect
|
diskfull install failed on env build linux lsxcatd v version git commit built thu may edt how to reproduce during diskfull provision url loading install kernel loading initial ramdisk suse linux zimage starting loaded at sp uncompressing elf header done bytes can t allocate memory for kernel image exit called ok
| 0
|
96,666
| 8,628,515,650
|
IssuesEvent
|
2018-11-21 17:36:56
|
pandas-dev/pandas
|
https://api.github.com/repos/pandas-dev/pandas
|
closed
|
TST: Test failure on 32bit for TestSparseGroupBy.test_aggfuncs
|
32bit Groupby Sparse Testing
|
The `test_aggfuncs` test in `/sparse/test_groupby.py` is failing on 32bit: https://travis-ci.org/MacPython/pandas-wheels/jobs/451706570#L5281
Specifically, the pytest output:
```
_______________________ TestSparseGroupBy.test_aggfuncs ________________________
self = <pandas.tests.sparse.test_groupby.TestSparseGroupBy object at 0xde3f08ec>
def test_aggfuncs(self):
sparse_grouped = self.sparse.groupby('A')
dense_grouped = self.dense.groupby('A')
result = sparse_grouped.mean().to_sparse()
expected = dense_grouped.mean().to_sparse()
> tm.assert_frame_equal(result, expected)
/venv/lib/python3.6/site-packages/pandas/tests/sparse/test_groupby.py:50:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/venv/lib/python3.6/site-packages/pandas/util/testing.py:1185: in assert_extension_array_equal
assert_numpy_array_equal(left_valid, right_valid)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
left = array([-0.17387645482451206, 0.3414148016424936], dtype=object)
right = array([-0.17387645482451206, 0.3414148016424937], dtype=object)
err_msg = None
def _raise(left, right, err_msg):
if err_msg is None:
if left.shape != right.shape:
raise_assert_detail(obj, '{obj} shapes are different'
.format(obj=obj), left.shape, right.shape)
diff = 0
for l, r in zip(left, right):
# count up differences
if not array_equivalent(l, r, strict_nan=strict_nan):
diff += 1
diff = diff * 100.0 / left.size
msg = '{obj} values are different ({pct} %)'.format(
obj=obj, pct=np.round(diff, 5))
> raise_assert_detail(obj, msg, left, right)
E AssertionError: numpy array are different
E
E numpy array values are different (50.0 %)
E [left]: [-0.17387645482451206, 0.3414148016424936]
E [right]: [-0.17387645482451206, 0.3414148016424937]
/venv/lib/python3.6/site-packages/pandas/util/testing.py:1146: AssertionError
```
The test in question:
https://github.com/pandas-dev/pandas/blob/efd1844daaadee29a57943597431611d554b6c4a/pandas/tests/sparse/test_groupby.py#L43-L48
xref #23318 : it looks like this is the pr that caused the change?
cc : @jreback @TomAugspurger @5hirish
|
1.0
|
TST: Test failure on 32bit for TestSparseGroupBy.test_aggfuncs - The `test_aggfuncs` test in `/sparse/test_groupby.py` is failing on 32bit: https://travis-ci.org/MacPython/pandas-wheels/jobs/451706570#L5281
Specifically, the pytest output:
```
_______________________ TestSparseGroupBy.test_aggfuncs ________________________
self = <pandas.tests.sparse.test_groupby.TestSparseGroupBy object at 0xde3f08ec>
def test_aggfuncs(self):
sparse_grouped = self.sparse.groupby('A')
dense_grouped = self.dense.groupby('A')
result = sparse_grouped.mean().to_sparse()
expected = dense_grouped.mean().to_sparse()
> tm.assert_frame_equal(result, expected)
/venv/lib/python3.6/site-packages/pandas/tests/sparse/test_groupby.py:50:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/venv/lib/python3.6/site-packages/pandas/util/testing.py:1185: in assert_extension_array_equal
assert_numpy_array_equal(left_valid, right_valid)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
left = array([-0.17387645482451206, 0.3414148016424936], dtype=object)
right = array([-0.17387645482451206, 0.3414148016424937], dtype=object)
err_msg = None
def _raise(left, right, err_msg):
if err_msg is None:
if left.shape != right.shape:
raise_assert_detail(obj, '{obj} shapes are different'
.format(obj=obj), left.shape, right.shape)
diff = 0
for l, r in zip(left, right):
# count up differences
if not array_equivalent(l, r, strict_nan=strict_nan):
diff += 1
diff = diff * 100.0 / left.size
msg = '{obj} values are different ({pct} %)'.format(
obj=obj, pct=np.round(diff, 5))
> raise_assert_detail(obj, msg, left, right)
E AssertionError: numpy array are different
E
E numpy array values are different (50.0 %)
E [left]: [-0.17387645482451206, 0.3414148016424936]
E [right]: [-0.17387645482451206, 0.3414148016424937]
/venv/lib/python3.6/site-packages/pandas/util/testing.py:1146: AssertionError
```
The test in question:
https://github.com/pandas-dev/pandas/blob/efd1844daaadee29a57943597431611d554b6c4a/pandas/tests/sparse/test_groupby.py#L43-L48
xref #23318 : it looks like this is the pr that caused the change?
cc : @jreback @TomAugspurger @5hirish
|
non_defect
|
tst test failure on for testsparsegroupby test aggfuncs the test aggfuncs test in sparse test groupby py is failing on specifically the pytest output testsparsegroupby test aggfuncs self def test aggfuncs self sparse grouped self sparse groupby a dense grouped self dense groupby a result sparse grouped mean to sparse expected dense grouped mean to sparse tm assert frame equal result expected venv lib site packages pandas tests sparse test groupby py venv lib site packages pandas util testing py in assert extension array equal assert numpy array equal left valid right valid left array dtype object right array dtype object err msg none def raise left right err msg if err msg is none if left shape right shape raise assert detail obj obj shapes are different format obj obj left shape right shape diff for l r in zip left right count up differences if not array equivalent l r strict nan strict nan diff diff diff left size msg obj values are different pct format obj obj pct np round diff raise assert detail obj msg left right e assertionerror numpy array are different e e numpy array values are different e e venv lib site packages pandas util testing py assertionerror the test in question xref it looks like this is the pr that caused the change cc jreback tomaugspurger
| 0
|
8,604
| 2,611,532,470
|
IssuesEvent
|
2015-02-27 06:03:41
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
Decreasing speed of aiming
|
auto-migrated Priority-Medium ReleaseBug-0.9.19 Type-Defect
|
```
What steps will reproduce the problem?
1. During the flight of a hog, notice the speed of aiming (it will be OK).
2. Now shot a rope one time, but not to cling to the surface.
3. Again notice the speed of aiming, it will be slower by about 30%.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Hedgewars 0.9.19
Archlinux with 3.9.2 kernel
Please provide any additional information below.
In the new version, I began to play shoppa worse, and I finally understood why.
```
Original issue reported on code.google.com by `srgl...@gmail.com` on 5 Jun 2013 at 1:13
|
1.0
|
Decreasing speed of aiming - ```
What steps will reproduce the problem?
1. During the flight of a hog, notice the speed of aiming (it will be OK).
2. Now shot a rope one time, but not to cling to the surface.
3. Again notice the speed of aiming, it will be slower by about 30%.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Hedgewars 0.9.19
Archlinux with 3.9.2 kernel
Please provide any additional information below.
In the new version, I began to play shoppa worse, and I finally understood why.
```
Original issue reported on code.google.com by `srgl...@gmail.com` on 5 Jun 2013 at 1:13
|
defect
|
decreasing speed of aiming what steps will reproduce the problem during the flight of a hog notice the speed of aiming it will be ok now shot a rope one time but not to cling to the surface again notice the speed of aiming it will be slower by about what is the expected output what do you see instead what version of the product are you using on what operating system hedgewars archlinux with kernel please provide any additional information below in the new version i began to play shoppa worse and i finally understood why original issue reported on code google com by srgl gmail com on jun at
| 1
|
5,378
| 2,610,186,513
|
IssuesEvent
|
2015-02-26 18:59:10
|
chrsmith/quchuseban
|
https://api.github.com/repos/chrsmith/quchuseban
|
opened
|
绿疗色斑怎么样才能去除
|
auto-migrated Priority-Medium Type-Defect
|
```
《摘要》
我不放弃依恋这样的冬日,喧闹的幸福,似乎已经遥远。曾��
�的她,至今已不再出现。我依旧不舍与这样的季节挥手。记�
��里,伤心的她,比这个季节更冰凉。独守破败的窗台,幻想
着她的经过。却迎来冬日暖阳,如昨日雪花,静静飘落,一��
�片轻轻柔柔的。一缕白烟,在我眼前升起,那是手中绿茶的�
��吸,也这样般轻柔。再美好也是曾经,脸上的痕迹提醒着我
,我们不会在相聚!色斑怎么样才能去除,
《客户案例》
林小姐 30岁<br>
没有色斑了,那里那个乐啊,看着现在白白嫩嫩的皮肤��
�我还真不敢相信还能拥有今天这样的美丽肌肤。<br>
想想以前,我喜欢化淡装,自认为随着年龄的增大,皮��
�变差了,为了让青春永驻容颜不老,经常更换不同品牌的化�
��品,听信了某个化妆品可以增白和改变肤质,用过一段时间
后,结果不但没改变,反面让我的面部出现了大量的色斑。��
�的我的脸一团糟。色斑聚集在脸颊两边,特别粗糙,还经常�
��皮,还有点发红,不过不是红血丝,很容易过敏。现在美白
不成又要祛斑,真是让人心烦啊。祛斑产品用了不少,激光��
�做过,但脸上的斑点就是不能彻底祛掉,还多了些黑印。这�
��症状持续了两年多的时间,如果不是遇到「黛芙薇尔精华液
」,我还真不知道要怎么祛除这可恶的斑点了。<br>
使用「黛芙薇尔精华液」半个多月后,我发现脸上的色��
�颜色就淡了许多,一个月后,斑块明显消褪了不少,皮肤也�
��慢变得白皙起来,脸色也变得红润有光泽了,精神状况也得
到明显的改善,两个周期服完后,简直像变了一个人似的,��
�肤变白了,而且脸上的色斑少了很多,就剩下鼻梁上面还有�
��些很巩固的斑点,看样子还非得再使用一个周期才能把这“
巩固分子”彻底的祛除了,于是我又订购了一个周期的产品��
�行巩固治疗。<br>
使用完之后,我脸上的色斑终于被我彻底搞定了,而且��
�肤也得到了很好的改善,变得白皙光滑了,看着现在的自己�
��感觉还年轻了不少呢。现在,我一般就用一些无香精,无化
学成分有洗面奶,保养我的皮肤,现在再也不去用什么化妆��
�了,脸上皮肤现在没有斑点,没有痘痘,再加上细腻嫩滑,�
��起来摸起来都不错。
阅读了色斑怎么样才能去除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎么样才能去除,同时为您分享祛斑小方法
选择适当的护肤品,不使用劣质化妆品,因其所含色素防腐��
�,能与汗水相混合,侵入皮肤内层,加速面部斑点的产生。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:56
|
1.0
|
绿疗色斑怎么样才能去除 - ```
《摘要》
我不放弃依恋这样的冬日,喧闹的幸福,似乎已经遥远。曾��
�的她,至今已不再出现。我依旧不舍与这样的季节挥手。记�
��里,伤心的她,比这个季节更冰凉。独守破败的窗台,幻想
着她的经过。却迎来冬日暖阳,如昨日雪花,静静飘落,一��
�片轻轻柔柔的。一缕白烟,在我眼前升起,那是手中绿茶的�
��吸,也这样般轻柔。再美好也是曾经,脸上的痕迹提醒着我
,我们不会在相聚!色斑怎么样才能去除,
《客户案例》
林小姐 30岁<br>
没有色斑了,那里那个乐啊,看着现在白白嫩嫩的皮肤��
�我还真不敢相信还能拥有今天这样的美丽肌肤。<br>
想想以前,我喜欢化淡装,自认为随着年龄的增大,皮��
�变差了,为了让青春永驻容颜不老,经常更换不同品牌的化�
��品,听信了某个化妆品可以增白和改变肤质,用过一段时间
后,结果不但没改变,反面让我的面部出现了大量的色斑。��
�的我的脸一团糟。色斑聚集在脸颊两边,特别粗糙,还经常�
��皮,还有点发红,不过不是红血丝,很容易过敏。现在美白
不成又要祛斑,真是让人心烦啊。祛斑产品用了不少,激光��
�做过,但脸上的斑点就是不能彻底祛掉,还多了些黑印。这�
��症状持续了两年多的时间,如果不是遇到「黛芙薇尔精华液
」,我还真不知道要怎么祛除这可恶的斑点了。<br>
使用「黛芙薇尔精华液」半个多月后,我发现脸上的色��
�颜色就淡了许多,一个月后,斑块明显消褪了不少,皮肤也�
��慢变得白皙起来,脸色也变得红润有光泽了,精神状况也得
到明显的改善,两个周期服完后,简直像变了一个人似的,��
�肤变白了,而且脸上的色斑少了很多,就剩下鼻梁上面还有�
��些很巩固的斑点,看样子还非得再使用一个周期才能把这“
巩固分子”彻底的祛除了,于是我又订购了一个周期的产品��
�行巩固治疗。<br>
使用完之后,我脸上的色斑终于被我彻底搞定了,而且��
�肤也得到了很好的改善,变得白皙光滑了,看着现在的自己�
��感觉还年轻了不少呢。现在,我一般就用一些无香精,无化
学成分有洗面奶,保养我的皮肤,现在再也不去用什么化妆��
�了,脸上皮肤现在没有斑点,没有痘痘,再加上细腻嫩滑,�
��起来摸起来都不错。
阅读了色斑怎么样才能去除,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎么样才能去除,同时为您分享祛斑小方法
选择适当的护肤品,不使用劣质化妆品,因其所含色素防腐��
�,能与汗水相混合,侵入皮肤内层,加速面部斑点的产生。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:56
|
defect
|
绿疗色斑怎么样才能去除 《摘要》 我不放弃依恋这样的冬日,喧闹的幸福,似乎已经遥远。曾�� �的她,至今已不再出现。我依旧不舍与这样的季节挥手。记� ��里,伤心的她,比这个季节更冰凉。独守破败的窗台,幻想 着她的经过。却迎来冬日暖阳,如昨日雪花,静静飘落,一�� �片轻轻柔柔的。一缕白烟,在我眼前升起,那是手中绿茶的� ��吸,也这样般轻柔。再美好也是曾经,脸上的痕迹提醒着我 ,我们不会在相聚!色斑怎么样才能去除, 《客户案例》 林小姐 没有色斑了,那里那个乐啊,看着现在白白嫩嫩的皮肤�� �我还真不敢相信还能拥有今天这样的美丽肌肤。 想想以前,我喜欢化淡装,自认为随着年龄的增大,皮�� �变差了,为了让青春永驻容颜不老,经常更换不同品牌的化� ��品,听信了某个化妆品可以增白和改变肤质,用过一段时间 后,结果不但没改变,反面让我的面部出现了大量的色斑。�� �的我的脸一团糟。色斑聚集在脸颊两边,特别粗糙,还经常� ��皮,还有点发红,不过不是红血丝,很容易过敏。现在美白 不成又要祛斑,真是让人心烦啊。祛斑产品用了不少,激光�� �做过,但脸上的斑点就是不能彻底祛掉,还多了些黑印。这� ��症状持续了两年多的时间,如果不是遇到「黛芙薇尔精华液 」,我还真不知道要怎么祛除这可恶的斑点了。 使用「黛芙薇尔精华液」半个多月后,我发现脸上的色�� �颜色就淡了许多,一个月后,斑块明显消褪了不少,皮肤也� ��慢变得白皙起来,脸色也变得红润有光泽了,精神状况也得 到明显的改善,两个周期服完后,简直像变了一个人似的,�� �肤变白了,而且脸上的色斑少了很多,就剩下鼻梁上面还有� ��些很巩固的斑点,看样子还非得再使用一个周期才能把这“ 巩固分子”彻底的祛除了,于是我又订购了一个周期的产品�� �行巩固治疗。 使用完之后,我脸上的色斑终于被我彻底搞定了,而且�� �肤也得到了很好的改善,变得白皙光滑了,看着现在的自己� ��感觉还年轻了不少呢。现在,我一般就用一些无香精,无化 学成分有洗面奶,保养我的皮肤,现在再也不去用什么化妆�� �了,脸上皮肤现在没有斑点,没有痘痘,再加上细腻嫩滑,� ��起来摸起来都不错。 阅读了色斑怎么样才能去除,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 色斑怎么样才能去除,同时为您分享祛斑小方法 选择适当的护肤品,不使用劣质化妆品,因其所含色素防腐�� �,能与汗水相混合,侵入皮肤内层,加速面部斑点的产生。 original issue reported on code google com by additive gmail com on jul at
| 1
|
63,777
| 3,198,652,529
|
IssuesEvent
|
2015-10-01 13:26:35
|
NRGI/resourcecontracts.org
|
https://api.github.com/repos/NRGI/resourcecontracts.org
|
closed
|
UI issue with Safari
|
bug Priority UI - Front page
|
Contracts not loadiing.
Console message:
[Error] TypeError: undefined is not a function (evaluating '$('select').select2({placeholder: "Select", allowClear: true, theme: "classic"})')
(anonymous function) (script.js, line 2)
j (jquery.min.js, line 2)
fireWith (jquery.min.js, line 2)
ready (jquery.min.js, line 2)
I (jquery.min.js, line 2)
[Warning] You are using the in-browser JSX transformer. Be sure to precompile your JSX for production - http://facebook.github.io/react/docs/tooling-integration.html#jsx (JSXTransformer.js, line 314)
[Error] SyntaxError: Unexpected token '('. Expected a ':' following the property name 'getDefaultProps'.
appendChild (view, line 23)
run (JSXTransformer.js, line 184)
check (JSXTransformer.js, line 238)
(anonymous function) (JSXTransformer.js, line 273)
onreadystatechange (JSXTransformer.js, line 208)
[Error] ReferenceError: Can't find variable: Waypoint
perform (react-with-addons.js, line 18419)
batchedUpdates (react-with-addons.js, line 9669)
enqueueUpdate (react-with-addons.js, line 16738)
enqueueUpdate (react-with-addons.js, line 16256)
enqueueForceUpdate (react-with-addons.js, line 16379)
forceUpdate (react-with-addons.js, line 6439)
(anonymous function) (view, line 142)
triggerEvents (backbone.js, line 353)
triggerApi (backbone.js, line 339)
eventsApi (backbone.js, line 137)
trigger (backbone.js, line 329)
reset (backbone.js, line 900)
success (backbone.js, line 991)
fire (jquery.js, line 1037)
fireWith (jquery.js, line 1148)
done (jquery.js, line 8074)
callback (jquery.js, line 8598)
|
1.0
|
UI issue with Safari - Contracts not loadiing.
Console message:
[Error] TypeError: undefined is not a function (evaluating '$('select').select2({placeholder: "Select", allowClear: true, theme: "classic"})')
(anonymous function) (script.js, line 2)
j (jquery.min.js, line 2)
fireWith (jquery.min.js, line 2)
ready (jquery.min.js, line 2)
I (jquery.min.js, line 2)
[Warning] You are using the in-browser JSX transformer. Be sure to precompile your JSX for production - http://facebook.github.io/react/docs/tooling-integration.html#jsx (JSXTransformer.js, line 314)
[Error] SyntaxError: Unexpected token '('. Expected a ':' following the property name 'getDefaultProps'.
appendChild (view, line 23)
run (JSXTransformer.js, line 184)
check (JSXTransformer.js, line 238)
(anonymous function) (JSXTransformer.js, line 273)
onreadystatechange (JSXTransformer.js, line 208)
[Error] ReferenceError: Can't find variable: Waypoint
perform (react-with-addons.js, line 18419)
batchedUpdates (react-with-addons.js, line 9669)
enqueueUpdate (react-with-addons.js, line 16738)
enqueueUpdate (react-with-addons.js, line 16256)
enqueueForceUpdate (react-with-addons.js, line 16379)
forceUpdate (react-with-addons.js, line 6439)
(anonymous function) (view, line 142)
triggerEvents (backbone.js, line 353)
triggerApi (backbone.js, line 339)
eventsApi (backbone.js, line 137)
trigger (backbone.js, line 329)
reset (backbone.js, line 900)
success (backbone.js, line 991)
fire (jquery.js, line 1037)
fireWith (jquery.js, line 1148)
done (jquery.js, line 8074)
callback (jquery.js, line 8598)
|
non_defect
|
ui issue with safari contracts not loadiing console message typeerror undefined is not a function evaluating select placeholder select allowclear true theme classic anonymous function script js line j jquery min js line firewith jquery min js line ready jquery min js line i jquery min js line you are using the in browser jsx transformer be sure to precompile your jsx for production jsxtransformer js line syntaxerror unexpected token expected a following the property name getdefaultprops appendchild view line run jsxtransformer js line check jsxtransformer js line anonymous function jsxtransformer js line onreadystatechange jsxtransformer js line referenceerror can t find variable waypoint perform react with addons js line batchedupdates react with addons js line enqueueupdate react with addons js line enqueueupdate react with addons js line enqueueforceupdate react with addons js line forceupdate react with addons js line anonymous function view line triggerevents backbone js line triggerapi backbone js line eventsapi backbone js line trigger backbone js line reset backbone js line success backbone js line fire jquery js line firewith jquery js line done jquery js line callback jquery js line
| 0
|
42,464
| 11,053,612,276
|
IssuesEvent
|
2019-12-10 11:45:49
|
ShaikASK/Testing
|
https://api.github.com/repos/ShaikASK/Testing
|
closed
|
Add "View Workflow" functionality in order to get the details of steps and document added in the workflow
|
Defect Fixed & TBD in B#56 HR Admin Module HR User Module P2 Workflow
|
Steps ;
1.Launch the URL
2.Sign in as HR user
3.Create Workflow and save it
4.Ensure that HR user is only having add & view workflow permission
5.Check the above created workflow
Experienced Behavior : Observed that user is able to view workflow thumbnail but he is not able to see the step description and document attached with workflow
Expected Behavior : Ensure that application should provide view workflow permission as like view new hire so that user should be able to view the details related to workflow
|
1.0
|
Add "View Workflow" functionality in order to get the details of steps and document added in the workflow - Steps ;
1.Launch the URL
2.Sign in as HR user
3.Create Workflow and save it
4.Ensure that HR user is only having add & view workflow permission
5.Check the above created workflow
Experienced Behavior : Observed that user is able to view workflow thumbnail but he is not able to see the step description and document attached with workflow
Expected Behavior : Ensure that application should provide view workflow permission as like view new hire so that user should be able to view the details related to workflow
|
defect
|
add view workflow functionality in order to get the details of steps and document added in the workflow steps launch the url sign in as hr user create workflow and save it ensure that hr user is only having add view workflow permission check the above created workflow experienced behavior observed that user is able to view workflow thumbnail but he is not able to see the step description and document attached with workflow expected behavior ensure that application should provide view workflow permission as like view new hire so that user should be able to view the details related to workflow
| 1
|
23,508
| 4,020,910,658
|
IssuesEvent
|
2016-05-16 20:07:08
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
appengine/taskqueue/pull-counter Needs Worker Tests
|
testing
|
The app in appengine/taskqueue/pull-counter has a worker that is run as a separate module using the worker.yaml. It listens on /_ah/start and then enters an infinite loop pulling from the pull taskqueue and pushing to datastore.
Currently the test just tests the post and addition to the queue, but not the worker function.
|
1.0
|
appengine/taskqueue/pull-counter Needs Worker Tests - The app in appengine/taskqueue/pull-counter has a worker that is run as a separate module using the worker.yaml. It listens on /_ah/start and then enters an infinite loop pulling from the pull taskqueue and pushing to datastore.
Currently the test just tests the post and addition to the queue, but not the worker function.
|
non_defect
|
appengine taskqueue pull counter needs worker tests the app in appengine taskqueue pull counter has a worker that is run as a separate module using the worker yaml it listens on ah start and then enters an infinite loop pulling from the pull taskqueue and pushing to datastore currently the test just tests the post and addition to the queue but not the worker function
| 0
|
240,192
| 18,294,997,035
|
IssuesEvent
|
2021-10-05 19:32:42
|
yjunechoe/ggtrace
|
https://api.github.com/repos/yjunechoe/ggtrace
|
closed
|
make trace_exprs argument optional
|
documentation enhancement priority: medium
|
If only `trace_steps` is provided, just run the specified steps
|
1.0
|
make trace_exprs argument optional - If only `trace_steps` is provided, just run the specified steps
|
non_defect
|
make trace exprs argument optional if only trace steps is provided just run the specified steps
| 0
|
824
| 10,592,138,162
|
IssuesEvent
|
2019-10-09 12:30:04
|
AzureAD/microsoft-authentication-library-for-dotnet
|
https://api.github.com/repos/AzureAD/microsoft-authentication-library-for-dotnet
|
closed
|
[Bug] MSAL should not stop developers from adding reserved scopes
|
ADFS In Progress Supportability bug
|
**Which Version of MSAL are you using ?**
MSAL 4.4
**Platform**
all
**What authentication flow has the issue?**
Client Credentials and all others.
- See discussion here: https://github.com/AzureAD/microsoft-authentication-library-for-dotnet/issues/1403
- Gist of it - In client credentials on ADFS, <resource>/.default does not work. Instead, users need to add "openid"
**Actual**: MSAL throws a client exception stating that users should not add their own reserved scope "openid"
**Expected**: MSAL should not throw, instead it should just merge scopes
|
True
|
[Bug] MSAL should not stop developers from adding reserved scopes - **Which Version of MSAL are you using ?**
MSAL 4.4
**Platform**
all
**What authentication flow has the issue?**
Client Credentials and all others.
- See discussion here: https://github.com/AzureAD/microsoft-authentication-library-for-dotnet/issues/1403
- Gist of it - In client credentials on ADFS, <resource>/.default does not work. Instead, users need to add "openid"
**Actual**: MSAL throws a client exception stating that users should not add their own reserved scope "openid"
**Expected**: MSAL should not throw, instead it should just merge scopes
|
non_defect
|
msal should not stop developers from adding reserved scopes which version of msal are you using msal platform all what authentication flow has the issue client credentials and all others see discussion here gist of it in client credentials on adfs default does not work instead users need to add openid actual msal throws a client exception stating that users should not add their own reserved scope openid expected msal should not throw instead it should just merge scopes
| 0
|
234,507
| 7,722,271,273
|
IssuesEvent
|
2018-05-24 08:46:09
|
sethballantyne/Game-Demos
|
https://api.github.com/repos/sethballantyne/Game-Demos
|
closed
|
Font constructor doesn't handle negative integer arguments.
|
Plexis-Game bug enhancement priority-low
|
the overloaded constructor Font(String ^, Surface ^, Size) doesn't handle the event that either member of Size may be a negative value; it should throw ArgumentException if either evaluate to < 0.
|
1.0
|
Font constructor doesn't handle negative integer arguments. - the overloaded constructor Font(String ^, Surface ^, Size) doesn't handle the event that either member of Size may be a negative value; it should throw ArgumentException if either evaluate to < 0.
|
non_defect
|
font constructor doesn t handle negative integer arguments the overloaded constructor font string surface size doesn t handle the event that either member of size may be a negative value it should throw argumentexception if either evaluate to
| 0
|
69,192
| 22,270,850,630
|
IssuesEvent
|
2022-06-10 12:12:40
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Diagram: not working properly when MaxConnections = 1
|
:lady_beetle: defect
|
### Describe the bug
Endpoint's Diagram is not working when maxConnections equals 1.
Checking the source code of DiagramRender.java, I checked that at method encodeEndpoint:
```
if (maxConnections != 1) {
wb.append(",maxConnections:").append(maxConnections);
}
```
If I change that condition to
```
if (maxConnections != 0) {
wb.append(",maxConnections:").append(maxConnections);
}
```
Then it'll work normally.
I also check on canvas object on JS side, that when maxConnection is set to 1, because of that piece of code, the maxConnections is not present, which makes the connection not work properly.
### Reproducer
1 - Create a diagram with no maxConnections set;
2 - Create a target endpoint with maxConnections = 1;
3 - Create a source endpoint with maxConnections = -1;
Try to connect both elements and it'll not work.
### Expected behavior
The source and target be able to connect
### PrimeFaces edition
Community
### PrimeFaces version
11.0.0
### Theme
_No response_
### JSF implementation
_No response_
### JSF version
2.3
### Browser(s)
Chrome
|
1.0
|
Diagram: not working properly when MaxConnections = 1 - ### Describe the bug
Endpoint's Diagram is not working when maxConnections equals 1.
Checking the source code of DiagramRender.java, I checked that at method encodeEndpoint:
```
if (maxConnections != 1) {
wb.append(",maxConnections:").append(maxConnections);
}
```
If I change that condition to
```
if (maxConnections != 0) {
wb.append(",maxConnections:").append(maxConnections);
}
```
Then it'll work normally.
I also check on canvas object on JS side, that when maxConnection is set to 1, because of that piece of code, the maxConnections is not present, which makes the connection not work properly.
### Reproducer
1 - Create a diagram with no maxConnections set;
2 - Create a target endpoint with maxConnections = 1;
3 - Create a source endpoint with maxConnections = -1;
Try to connect both elements and it'll not work.
### Expected behavior
The source and target be able to connect
### PrimeFaces edition
Community
### PrimeFaces version
11.0.0
### Theme
_No response_
### JSF implementation
_No response_
### JSF version
2.3
### Browser(s)
Chrome
|
defect
|
diagram not working properly when maxconnections describe the bug endpoint s diagram is not working when maxconnections equals checking the source code of diagramrender java i checked that at method encodeendpoint if maxconnections wb append maxconnections append maxconnections if i change that condition to if maxconnections wb append maxconnections append maxconnections then it ll work normally i also check on canvas object on js side that when maxconnection is set to because of that piece of code the maxconnections is not present which makes the connection not work properly reproducer create a diagram with no maxconnections set create a target endpoint with maxconnections create a source endpoint with maxconnections try to connect both elements and it ll not work expected behavior the source and target be able to connect primefaces edition community primefaces version theme no response jsf implementation no response jsf version browser s chrome
| 1
|
227,694
| 25,107,722,189
|
IssuesEvent
|
2022-11-08 17:52:54
|
bitbar/android-gradle-plugin
|
https://api.github.com/repos/bitbar/android-gradle-plugin
|
closed
|
CVE-2022-42889 (High) detected in commons-text-1.9.jar - autoclosed
|
security vulnerability
|
## CVE-2022-42889 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-text-1.9.jar</b></p></summary>
<p>Apache Commons Text is a library focused on algorithms working on strings.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-text">https://commons.apache.org/proper/commons-text</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-text/1.9/ba6ac8c2807490944a0a27f6f8e68fb5ed2e80e2/commons-text-1.9.jar</p>
<p>
Dependency Hierarchy:
- testdroid-api-3.1.jar (Root Library)
- :x: **commons-text-1.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bitbar/android-gradle-plugin/commit/e89d66259a7e239dd3b023916815ce42b084b121">e89d66259a7e239dd3b023916815ce42b084b121</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons Text performs variable interpolation, allowing properties to be dynamically evaluated and expanded. The standard format for interpolation is "${prefix:name}", where "prefix" is used to locate an instance of org.apache.commons.text.lookup.StringLookup that performs the interpolation. Starting with version 1.5 and continuing through 1.9, the set of default Lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers. These lookups are: - "script" - execute expressions using the JVM script execution engine (javax.script) - "dns" - resolve dns records - "url" - load values from urls, including from remote servers Applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used. Users are recommended to upgrade to Apache Commons Text 1.10.0, which disables the problematic interpolators by default.
<p>Publish Date: 2022-10-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-42889>CVE-2022-42889</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2022/10/13/4">https://www.openwall.com/lists/oss-security/2022/10/13/4</a></p>
<p>Release Date: 2022-10-13</p>
<p>Fix Resolution (org.apache.commons:commons-text): 1.10.0</p>
<p>Direct dependency fix Resolution (com.testdroid:testdroid-api): 3.7</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
True
|
CVE-2022-42889 (High) detected in commons-text-1.9.jar - autoclosed - ## CVE-2022-42889 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-text-1.9.jar</b></p></summary>
<p>Apache Commons Text is a library focused on algorithms working on strings.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-text">https://commons.apache.org/proper/commons-text</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-text/1.9/ba6ac8c2807490944a0a27f6f8e68fb5ed2e80e2/commons-text-1.9.jar</p>
<p>
Dependency Hierarchy:
- testdroid-api-3.1.jar (Root Library)
- :x: **commons-text-1.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bitbar/android-gradle-plugin/commit/e89d66259a7e239dd3b023916815ce42b084b121">e89d66259a7e239dd3b023916815ce42b084b121</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons Text performs variable interpolation, allowing properties to be dynamically evaluated and expanded. The standard format for interpolation is "${prefix:name}", where "prefix" is used to locate an instance of org.apache.commons.text.lookup.StringLookup that performs the interpolation. Starting with version 1.5 and continuing through 1.9, the set of default Lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers. These lookups are: - "script" - execute expressions using the JVM script execution engine (javax.script) - "dns" - resolve dns records - "url" - load values from urls, including from remote servers Applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used. Users are recommended to upgrade to Apache Commons Text 1.10.0, which disables the problematic interpolators by default.
<p>Publish Date: 2022-10-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-42889>CVE-2022-42889</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2022/10/13/4">https://www.openwall.com/lists/oss-security/2022/10/13/4</a></p>
<p>Release Date: 2022-10-13</p>
<p>Fix Resolution (org.apache.commons:commons-text): 1.10.0</p>
<p>Direct dependency fix Resolution (com.testdroid:testdroid-api): 3.7</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
|
non_defect
|
cve high detected in commons text jar autoclosed cve high severity vulnerability vulnerable library commons text jar apache commons text is a library focused on algorithms working on strings library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org apache commons commons text commons text jar dependency hierarchy testdroid api jar root library x commons text jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons text performs variable interpolation allowing properties to be dynamically evaluated and expanded the standard format for interpolation is prefix name where prefix is used to locate an instance of org apache commons text lookup stringlookup that performs the interpolation starting with version and continuing through the set of default lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers these lookups are script execute expressions using the jvm script execution engine javax script dns resolve dns records url load values from urls including from remote servers applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used users are recommended to upgrade to apache commons text which disables the problematic interpolators by default publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache commons commons text direct dependency fix resolution com testdroid testdroid api rescue worker helmet automatic remediation is available for this issue
| 0
|
20,609
| 3,388,318,577
|
IssuesEvent
|
2015-11-29 06:22:31
|
netty/netty
|
https://api.github.com/repos/netty/netty
|
closed
|
IllegalReferenceCountException from HttpObjectDecoder when switching a protocol
|
defect
|
I found `HttpObjectDecoder.decode()` raises an `IllegalReferenceCountException` at the `case UPGRADED` block. After pulling my hairs, I found that it can happen when a user removes the `HttpObjectDecoder` from the pipeline via `HttpClientUpgradeHandler`. The following buffer access trace should be a good hint:
```
Recent access records: 5
#5:
io.netty.buffer.AdvancedLeakAwareByteBuf.readBytes(AdvancedLeakAwareByteBuf.java:434)
io.netty.handler.codec.ByteToMessageDecoder.handlerRemoved(ByteToMessageDecoder.java:214)
io.netty.channel.DefaultChannelPipeline.callHandlerRemoved0(DefaultChannelPipeline.java:629)
io.netty.channel.DefaultChannelPipeline.callHandlerRemoved(DefaultChannelPipeline.java:623)
io.netty.channel.DefaultChannelPipeline.remove0(DefaultChannelPipeline.java:452)
io.netty.channel.DefaultChannelPipeline.remove(DefaultChannelPipeline.java:423)
io.netty.channel.DefaultChannelPipeline.remove(DefaultChannelPipeline.java:412)
io.netty.handler.codec.http.HttpClientCodec.upgradeFrom(HttpClientCodec.java:95)
io.netty.handler.codec.http.HttpClientUpgradeHandler.decode(HttpClientUpgradeHandler.java:230)
io.netty.handler.codec.http.HttpClientUpgradeHandler.decode(HttpClientUpgradeHandler.java:38)
io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:89)
io.netty.channel.ChannelHandlerInvokerUtil.invokeChannelReadNow(ChannelHandlerInvokerUtil.java:83)
io.netty.channel.DefaultChannelHandlerInvoker.invokeChannelRead(DefaultChannelHandlerInvoker.java:163)
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:155)
io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:276)
io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:354)
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244)
io.netty.channel.ChannelHandlerInvokerUtil.invokeChannelReadNow(ChannelHandlerInvokerUtil.java:83)
io.netty.channel.DefaultChannelHandlerInvoker.invokeChannelRead(DefaultChannelHandlerInvoker.java:163)
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:155)
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:950)
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:125)
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:510)
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:467)
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:381)
io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:353)
io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137)
java.lang.Thread.run(Thread.java:745)
```
It basically means:
1. HttpObjectDecoder decodes the upgrade response from the server. The HttpObjectDecoder enters the UPGRADED state
2. HttpClientUpgradeHandler upgrades the protocol and removes the HttpObjectDecoder from the pipeline. HttpObjectDecoder.handlerRemoved() releases the buffer, but we are still at HttpObjectDecoder.decode().
3. A user gets an IllegalReferenceCountException.
|
1.0
|
IllegalReferenceCountException from HttpObjectDecoder when switching a protocol - I found `HttpObjectDecoder.decode()` raises an `IllegalReferenceCountException` at the `case UPGRADED` block. After pulling my hairs, I found that it can happen when a user removes the `HttpObjectDecoder` from the pipeline via `HttpClientUpgradeHandler`. The following buffer access trace should be a good hint:
```
Recent access records: 5
#5:
io.netty.buffer.AdvancedLeakAwareByteBuf.readBytes(AdvancedLeakAwareByteBuf.java:434)
io.netty.handler.codec.ByteToMessageDecoder.handlerRemoved(ByteToMessageDecoder.java:214)
io.netty.channel.DefaultChannelPipeline.callHandlerRemoved0(DefaultChannelPipeline.java:629)
io.netty.channel.DefaultChannelPipeline.callHandlerRemoved(DefaultChannelPipeline.java:623)
io.netty.channel.DefaultChannelPipeline.remove0(DefaultChannelPipeline.java:452)
io.netty.channel.DefaultChannelPipeline.remove(DefaultChannelPipeline.java:423)
io.netty.channel.DefaultChannelPipeline.remove(DefaultChannelPipeline.java:412)
io.netty.handler.codec.http.HttpClientCodec.upgradeFrom(HttpClientCodec.java:95)
io.netty.handler.codec.http.HttpClientUpgradeHandler.decode(HttpClientUpgradeHandler.java:230)
io.netty.handler.codec.http.HttpClientUpgradeHandler.decode(HttpClientUpgradeHandler.java:38)
io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:89)
io.netty.channel.ChannelHandlerInvokerUtil.invokeChannelReadNow(ChannelHandlerInvokerUtil.java:83)
io.netty.channel.DefaultChannelHandlerInvoker.invokeChannelRead(DefaultChannelHandlerInvoker.java:163)
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:155)
io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:276)
io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:354)
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:244)
io.netty.channel.ChannelHandlerInvokerUtil.invokeChannelReadNow(ChannelHandlerInvokerUtil.java:83)
io.netty.channel.DefaultChannelHandlerInvoker.invokeChannelRead(DefaultChannelHandlerInvoker.java:163)
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:155)
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:950)
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:125)
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:510)
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:467)
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:381)
io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:353)
io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137)
java.lang.Thread.run(Thread.java:745)
```
It basically means:
1. HttpObjectDecoder decodes the upgrade response from the server. The HttpObjectDecoder enters the UPGRADED state
2. HttpClientUpgradeHandler upgrades the protocol and removes the HttpObjectDecoder from the pipeline. HttpObjectDecoder.handlerRemoved() releases the buffer, but we are still at HttpObjectDecoder.decode().
3. A user gets an IllegalReferenceCountException.
|
defect
|
illegalreferencecountexception from httpobjectdecoder when switching a protocol i found httpobjectdecoder decode raises an illegalreferencecountexception at the case upgraded block after pulling my hairs i found that it can happen when a user removes the httpobjectdecoder from the pipeline via httpclientupgradehandler the following buffer access trace should be a good hint recent access records io netty buffer advancedleakawarebytebuf readbytes advancedleakawarebytebuf java io netty handler codec bytetomessagedecoder handlerremoved bytetomessagedecoder java io netty channel defaultchannelpipeline defaultchannelpipeline java io netty channel defaultchannelpipeline callhandlerremoved defaultchannelpipeline java io netty channel defaultchannelpipeline defaultchannelpipeline java io netty channel defaultchannelpipeline remove defaultchannelpipeline java io netty channel defaultchannelpipeline remove defaultchannelpipeline java io netty handler codec http httpclientcodec upgradefrom httpclientcodec java io netty handler codec http httpclientupgradehandler decode httpclientupgradehandler java io netty handler codec http httpclientupgradehandler decode httpclientupgradehandler java io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java io netty channel channelhandlerinvokerutil invokechannelreadnow channelhandlerinvokerutil java io netty channel defaultchannelhandlerinvoker invokechannelread defaultchannelhandlerinvoker java io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java io netty channel channelhandlerinvokerutil invokechannelreadnow channelhandlerinvokerutil java io netty channel defaultchannelhandlerinvoker invokechannelread defaultchannelhandlerinvoker java io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java io netty channel nio nioeventloop processselectedkey nioeventloop java io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java io netty channel nio nioeventloop processselectedkeys nioeventloop java io netty channel nio nioeventloop run nioeventloop java io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java io netty util concurrent defaultthreadfactory defaultrunnabledecorator run defaultthreadfactory java java lang thread run thread java it basically means httpobjectdecoder decodes the upgrade response from the server the httpobjectdecoder enters the upgraded state httpclientupgradehandler upgrades the protocol and removes the httpobjectdecoder from the pipeline httpobjectdecoder handlerremoved releases the buffer but we are still at httpobjectdecoder decode a user gets an illegalreferencecountexception
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.