Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
718,517
| 24,720,620,872
|
IssuesEvent
|
2022-10-20 10:22:07
|
YangCatalog/backend
|
https://api.github.com/repos/YangCatalog/backend
|
closed
|
Checkbox to disable synonyms
|
enhancement Priority: Low
|
Having the synonyms is really great. But in some cases a user may not want to use it, so having a checkbox to disable it would be great.
Also when synonyms are used, it'd be good for the user to know what synonyms were used.
|
1.0
|
Checkbox to disable synonyms - Having the synonyms is really great. But in some cases a user may not want to use it, so having a checkbox to disable it would be great.
Also when synonyms are used, it'd be good for the user to know what synonyms were used.
|
non_process
|
checkbox to disable synonyms having the synonyms is really great but in some cases a user may not want to use it so having a checkbox to disable it would be great also when synonyms are used it d be good for the user to know what synonyms were used
| 0
|
119,035
| 15,393,564,869
|
IssuesEvent
|
2021-03-03 16:51:58
|
grommet/hpe-design-system
|
https://api.github.com/repos/grommet/hpe-design-system
|
closed
|
Update Figma file to demonstrate drop suggestions on Search
|
Implement design
|
<img width="442" alt="Screen Shot 2021-02-03 at 10 44 52 AM" src="https://user-images.githubusercontent.com/62117454/106793917-ed313d80-660c-11eb-8819-28f6884ace7e.png">
Please update Figma files to align with DS site [(https://design-system.hpe.design/components/search)](url)
|
1.0
|
Update Figma file to demonstrate drop suggestions on Search - <img width="442" alt="Screen Shot 2021-02-03 at 10 44 52 AM" src="https://user-images.githubusercontent.com/62117454/106793917-ed313d80-660c-11eb-8819-28f6884ace7e.png">
Please update Figma files to align with DS site [(https://design-system.hpe.design/components/search)](url)
|
non_process
|
update figma file to demonstrate drop suggestions on search img width alt screen shot at am src please update figma files to align with ds site url
| 0
|
381,413
| 11,276,538,693
|
IssuesEvent
|
2020-01-14 23:33:17
|
googleapis/google-api-java-client-services
|
https://api.github.com/repos/googleapis/google-api-java-client-services
|
closed
|
Synthesis failed for acceleratedmobilepageurl
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate acceleratedmobilepageurl. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Checking out files: 24% (15878/65361)
Checking out files: 25% (16341/65361)
Checking out files: 26% (16994/65361)
Checking out files: 27% (17648/65361)
Checking out files: 28% (18302/65361)
Checking out files: 29% (18955/65361)
Checking out files: 30% (19609/65361)
Checking out files: 31% (20262/65361)
Checking out files: 32% (20916/65361)
Checking out files: 33% (21570/65361)
Checking out files: 34% (22223/65361)
Checking out files: 35% (22877/65361)
Checking out files: 36% (23530/65361)
Checking out files: 37% (24184/65361)
Checking out files: 38% (24838/65361)
Checking out files: 39% (25491/65361)
Checking out files: 40% (26145/65361)
Checking out files: 41% (26799/65361)
Checking out files: 42% (27452/65361)
Checking out files: 43% (28106/65361)
Checking out files: 44% (28759/65361)
Checking out files: 45% (29413/65361)
Checking out files: 46% (30067/65361)
Checking out files: 47% (30720/65361)
Checking out files: 48% (31374/65361)
Checking out files: 49% (32027/65361)
Checking out files: 50% (32681/65361)
Checking out files: 51% (33335/65361)
Checking out files: 52% (33988/65361)
Checking out files: 53% (34642/65361)
Checking out files: 53% (35269/65361)
Checking out files: 54% (35295/65361)
Checking out files: 55% (35949/65361)
Checking out files: 56% (36603/65361)
Checking out files: 57% (37256/65361)
Checking out files: 58% (37910/65361)
Checking out files: 59% (38563/65361)
Checking out files: 60% (39217/65361)
Checking out files: 61% (39871/65361)
Checking out files: 62% (40524/65361)
Checking out files: 63% (41178/65361)
Checking out files: 64% (41832/65361)
Checking out files: 65% (42485/65361)
Checking out files: 66% (43139/65361)
Checking out files: 67% (43792/65361)
Checking out files: 68% (44446/65361)
Checking out files: 69% (45100/65361)
Checking out files: 70% (45753/65361)
Checking out files: 71% (46407/65361)
Checking out files: 72% (47060/65361)
Checking out files: 73% (47714/65361)
Checking out files: 74% (48368/65361)
Checking out files: 75% (49021/65361)
Checking out files: 76% (49675/65361)
Checking out files: 77% (50328/65361)
Checking out files: 78% (50982/65361)
Checking out files: 79% (51636/65361)
Checking out files: 80% (52289/65361)
Checking out files: 81% (52943/65361)
Checking out files: 82% (53597/65361)
Checking out files: 82% (53676/65361)
Checking out files: 83% (54250/65361)
Checking out files: 84% (54904/65361)
Checking out files: 85% (55557/65361)
Checking out files: 86% (56211/65361)
Checking out files: 87% (56865/65361)
Checking out files: 88% (57518/65361)
Checking out files: 89% (58172/65361)
Checking out files: 90% (58825/65361)
Checking out files: 91% (59479/65361)
Checking out files: 92% (60133/65361)
Checking out files: 93% (60786/65361)
Checking out files: 94% (61440/65361)
Checking out files: 95% (62093/65361)
Checking out files: 96% (62747/65361)
Checking out files: 97% (63401/65361)
Checking out files: 98% (64054/65361)
Checking out files: 99% (64708/65361)
Checking out files: 100% (65361/65361)
Checking out files: 100% (65361/65361), done.
Switched to branch 'autosynth-acceleratedmobilepageurl'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
|
1.0
|
Synthesis failed for acceleratedmobilepageurl - Hello! Autosynth couldn't regenerate acceleratedmobilepageurl. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Checking out files: 24% (15878/65361)
Checking out files: 25% (16341/65361)
Checking out files: 26% (16994/65361)
Checking out files: 27% (17648/65361)
Checking out files: 28% (18302/65361)
Checking out files: 29% (18955/65361)
Checking out files: 30% (19609/65361)
Checking out files: 31% (20262/65361)
Checking out files: 32% (20916/65361)
Checking out files: 33% (21570/65361)
Checking out files: 34% (22223/65361)
Checking out files: 35% (22877/65361)
Checking out files: 36% (23530/65361)
Checking out files: 37% (24184/65361)
Checking out files: 38% (24838/65361)
Checking out files: 39% (25491/65361)
Checking out files: 40% (26145/65361)
Checking out files: 41% (26799/65361)
Checking out files: 42% (27452/65361)
Checking out files: 43% (28106/65361)
Checking out files: 44% (28759/65361)
Checking out files: 45% (29413/65361)
Checking out files: 46% (30067/65361)
Checking out files: 47% (30720/65361)
Checking out files: 48% (31374/65361)
Checking out files: 49% (32027/65361)
Checking out files: 50% (32681/65361)
Checking out files: 51% (33335/65361)
Checking out files: 52% (33988/65361)
Checking out files: 53% (34642/65361)
Checking out files: 53% (35269/65361)
Checking out files: 54% (35295/65361)
Checking out files: 55% (35949/65361)
Checking out files: 56% (36603/65361)
Checking out files: 57% (37256/65361)
Checking out files: 58% (37910/65361)
Checking out files: 59% (38563/65361)
Checking out files: 60% (39217/65361)
Checking out files: 61% (39871/65361)
Checking out files: 62% (40524/65361)
Checking out files: 63% (41178/65361)
Checking out files: 64% (41832/65361)
Checking out files: 65% (42485/65361)
Checking out files: 66% (43139/65361)
Checking out files: 67% (43792/65361)
Checking out files: 68% (44446/65361)
Checking out files: 69% (45100/65361)
Checking out files: 70% (45753/65361)
Checking out files: 71% (46407/65361)
Checking out files: 72% (47060/65361)
Checking out files: 73% (47714/65361)
Checking out files: 74% (48368/65361)
Checking out files: 75% (49021/65361)
Checking out files: 76% (49675/65361)
Checking out files: 77% (50328/65361)
Checking out files: 78% (50982/65361)
Checking out files: 79% (51636/65361)
Checking out files: 80% (52289/65361)
Checking out files: 81% (52943/65361)
Checking out files: 82% (53597/65361)
Checking out files: 82% (53676/65361)
Checking out files: 83% (54250/65361)
Checking out files: 84% (54904/65361)
Checking out files: 85% (55557/65361)
Checking out files: 86% (56211/65361)
Checking out files: 87% (56865/65361)
Checking out files: 88% (57518/65361)
Checking out files: 89% (58172/65361)
Checking out files: 90% (58825/65361)
Checking out files: 91% (59479/65361)
Checking out files: 92% (60133/65361)
Checking out files: 93% (60786/65361)
Checking out files: 94% (61440/65361)
Checking out files: 95% (62093/65361)
Checking out files: 96% (62747/65361)
Checking out files: 97% (63401/65361)
Checking out files: 98% (64054/65361)
Checking out files: 99% (64708/65361)
Checking out files: 100% (65361/65361)
Checking out files: 100% (65361/65361), done.
Switched to branch 'autosynth-acceleratedmobilepageurl'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module>
main()
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main
last_synth_commit_hash = get_last_metadata_commit(args.metadata_path)
File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit
text=True,
File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run
with Popen(*popenargs, **kwargs) as process:
TypeError: __init__() got an unexpected keyword argument 'text'
```
Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
|
non_process
|
synthesis failed for acceleratedmobilepageurl hello autosynth couldn t regenerate acceleratedmobilepageurl broken heart here s the output from running synth py cloning into working repo checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files done switched to branch autosynth acceleratedmobilepageurl traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log
| 0
|
14,041
| 16,849,514,532
|
IssuesEvent
|
2021-06-20 07:54:41
|
log2timeline/plaso
|
https://api.github.com/repos/log2timeline/plaso
|
opened
|
Add support for preprocessing plugins to indicate they dependent on another plugin
|
enhancement preprocessing
|
To resolve Windows localized time zone names WindowsAvailableTimeZones needs to run before WindowsTimezone (https://github.com/log2timeline/plaso/issues/2673). Change preprocessing plugins to define such dependencies.
|
1.0
|
Add support for preprocessing plugins to indicate they dependent on another plugin - To resolve Windows localized time zone names WindowsAvailableTimeZones needs to run before WindowsTimezone (https://github.com/log2timeline/plaso/issues/2673). Change preprocessing plugins to define such dependencies.
|
process
|
add support for preprocessing plugins to indicate they dependent on another plugin to resolve windows localized time zone names windowsavailabletimezones needs to run before windowstimezone change preprocessing plugins to define such dependencies
| 1
|
14,804
| 18,104,027,647
|
IssuesEvent
|
2021-09-22 17:06:15
|
GoogleCloudPlatform/php-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/php-docs-samples
|
closed
|
storage: separate each sample into a separate file
|
priority: p2 type: process help wanted api: storage samples
|
Each sample should be in a standalone file. This workitem is to track that work.
|
1.0
|
storage: separate each sample into a separate file - Each sample should be in a standalone file. This workitem is to track that work.
|
process
|
storage separate each sample into a separate file each sample should be in a standalone file this workitem is to track that work
| 1
|
12,815
| 15,190,014,434
|
IssuesEvent
|
2021-02-15 17:14:48
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
Process does not run successfully on HP UX
|
Bug Process Status: Needs Review
|
**Symfony version(s) affected**: Symfony\Process 2.8
**Description**
NOTE: This is running on HP UX B11.31 which may not be worth your time as it's about to be discontinued. Still, this is my poison.
I'm using Gitlist which depends on gitter which depends on Symfony\Process. Process is used to run a git log command, and while I can run this command in the commandline without errors, it does not actually run using Symfony\Process. Or at least `isSuccessful()` returns `false`.
Other git commands run fine.
**How to reproduce**
```php
$process = new Process("'/usr/local/bin/git' -c \"color.ui\"=false log --skip=0 --max-count=15 --pretty=format:\"%H%h%T%P%aN%aE%at%cN%cE%ct\" master", "/sw/git/des_mnu");
$process->setTimeout(180);
$process->run();
if (!$process->isSuccessful()) {
die("Rattle");
}
```
This always fails.
Other commands, like
```
'/usr/local/bin/git' -c \"color.ui\"=false branch
```
run fine.
Note that running this command just gives multiple plain-text lines with commit info. Not sure why it is failing.
**Possible Solution**
No real idea.
**Additional context**
OS: An antiquated HP UX B11.31
|
1.0
|
Process does not run successfully on HP UX - **Symfony version(s) affected**: Symfony\Process 2.8
**Description**
NOTE: This is running on HP UX B11.31 which may not be worth your time as it's about to be discontinued. Still, this is my poison.
I'm using Gitlist which depends on gitter which depends on Symfony\Process. Process is used to run a git log command, and while I can run this command in the commandline without errors, it does not actually run using Symfony\Process. Or at least `isSuccessful()` returns `false`.
Other git commands run fine.
**How to reproduce**
```php
$process = new Process("'/usr/local/bin/git' -c \"color.ui\"=false log --skip=0 --max-count=15 --pretty=format:\"%H%h%T%P%aN%aE%at%cN%cE%ct\" master", "/sw/git/des_mnu");
$process->setTimeout(180);
$process->run();
if (!$process->isSuccessful()) {
die("Rattle");
}
```
This always fails.
Other commands, like
```
'/usr/local/bin/git' -c \"color.ui\"=false branch
```
run fine.
Note that running this command just gives multiple plain-text lines with commit info. Not sure why it is failing.
**Possible Solution**
No real idea.
**Additional context**
OS: An antiquated HP UX B11.31
|
process
|
process does not run successfully on hp ux symfony version s affected symfony process description note this is running on hp ux which may not be worth your time as it s about to be discontinued still this is my poison i m using gitlist which depends on gitter which depends on symfony process process is used to run a git log command and while i can run this command in the commandline without errors it does not actually run using symfony process or at least issuccessful returns false other git commands run fine how to reproduce php process new process usr local bin git c color ui false log skip max count pretty format h h t p an ae at cn ce ct master sw git des mnu process settimeout process run if process issuccessful die rattle this always fails other commands like usr local bin git c color ui false branch run fine note that running this command just gives multiple plain text lines with commit info not sure why it is failing possible solution no real idea additional context os an antiquated hp ux
| 1
|
35,585
| 17,140,578,731
|
IssuesEvent
|
2021-07-13 09:09:46
|
Yoast/wordpress-seo
|
https://api.github.com/repos/Yoast/wordpress-seo
|
closed
|
fill_cache creates high server load
|
Yoast: Management component: indexables component: performance severity: minor
|
<!-- Please use this template when creating an issue.
- Please check the boxes after you've created your issue.
- Please use the latest version of Yoast SEO.-->
* [x] I've read and understood the [contribution guidelines](https://github.com/Yoast/wordpress-seo/blob/trunk/.github/CONTRIBUTING.md).
* [x] I've searched for any related issues and avoided creating a duplicate issue.
### Please give us a description of what happened.
After upgrade from Yoast SEO from 13.5 to 16.0.2 experiencing high server load.
### Please describe what you expected to happen and why.
The reason is very slow SQL query being executed on every ajax request:
```
SELECT SQL_CALC_FOUND_ROWS wp_posts.ID FROM wp_posts WHERE 1=1 AND wp_posts.post_type = 'post' AND (wp_posts.post_status = 'publish' OR wp_posts.post_status = 'future' OR wp_posts.post_status = 'draft' OR wp_posts.post_status = 'pending') ORDER BY wp_posts.post_date DESC LIMIT 0, 10
```
This query takes 0.5 sec on our server and EXPLAIN shows:
```
+------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+
| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra |
+------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+
| 1 | SIMPLE | wp_posts | range | type_status_date | type_status_date | 164 | NULL | 453574 | Using where; Using index; Using filesort |
+------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+
```
It is being called from:
```
do_action('admin_init'), WP_Hook->do_action, WP_Hook->apply_filters, Yoast\WP\SEO\Integrations\Admin\Admin_Columns_Cache_Integration->fill_cache, WP_Query->get_posts
```
Temporary workaround is to comment out fill_cache call in function register_hooks() in wp-content/plugins/wordpress-seo/src/integrations/admin/admin-columns-cache-integration.php:
```
// \add_action( 'admin_init', [ $this, 'fill_cache' ] );
```
but actually this line should be checked:
```
$posts = empty( $wp_query->posts ) ? $wp_query->get_posts() : $wp_query->posts;
```
#### Used versions
* WordPress version: 5.7
* Yoast SEO version: 16.0.2
|
True
|
fill_cache creates high server load - <!-- Please use this template when creating an issue.
- Please check the boxes after you've created your issue.
- Please use the latest version of Yoast SEO.-->
* [x] I've read and understood the [contribution guidelines](https://github.com/Yoast/wordpress-seo/blob/trunk/.github/CONTRIBUTING.md).
* [x] I've searched for any related issues and avoided creating a duplicate issue.
### Please give us a description of what happened.
After upgrade from Yoast SEO from 13.5 to 16.0.2 experiencing high server load.
### Please describe what you expected to happen and why.
The reason is very slow SQL query being executed on every ajax request:
```
SELECT SQL_CALC_FOUND_ROWS wp_posts.ID FROM wp_posts WHERE 1=1 AND wp_posts.post_type = 'post' AND (wp_posts.post_status = 'publish' OR wp_posts.post_status = 'future' OR wp_posts.post_status = 'draft' OR wp_posts.post_status = 'pending') ORDER BY wp_posts.post_date DESC LIMIT 0, 10
```
This query takes 0.5 sec on our server and EXPLAIN shows:
```
+------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+
| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra |
+------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+
| 1 | SIMPLE | wp_posts | range | type_status_date | type_status_date | 164 | NULL | 453574 | Using where; Using index; Using filesort |
+------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+
```
It is being called from:
```
do_action('admin_init'), WP_Hook->do_action, WP_Hook->apply_filters, Yoast\WP\SEO\Integrations\Admin\Admin_Columns_Cache_Integration->fill_cache, WP_Query->get_posts
```
Temporary workaround is to comment out fill_cache call in function register_hooks() in wp-content/plugins/wordpress-seo/src/integrations/admin/admin-columns-cache-integration.php:
```
// \add_action( 'admin_init', [ $this, 'fill_cache' ] );
```
but actually this line should be checked:
```
$posts = empty( $wp_query->posts ) ? $wp_query->get_posts() : $wp_query->posts;
```
#### Used versions
* WordPress version: 5.7
* Yoast SEO version: 16.0.2
|
non_process
|
fill cache creates high server load please use this template when creating an issue please check the boxes after you ve created your issue please use the latest version of yoast seo i ve read and understood the i ve searched for any related issues and avoided creating a duplicate issue please give us a description of what happened after upgrade from yoast seo from to experiencing high server load please describe what you expected to happen and why the reason is very slow sql query being executed on every ajax request select sql calc found rows wp posts id from wp posts where and wp posts post type post and wp posts post status publish or wp posts post status future or wp posts post status draft or wp posts post status pending order by wp posts post date desc limit this query takes sec on our server and explain shows id select type table type possible keys key key len ref rows extra simple wp posts range type status date type status date null using where using index using filesort it is being called from do action admin init wp hook do action wp hook apply filters yoast wp seo integrations admin admin columns cache integration fill cache wp query get posts temporary workaround is to comment out fill cache call in function register hooks in wp content plugins wordpress seo src integrations admin admin columns cache integration php add action admin init but actually this line should be checked posts empty wp query posts wp query get posts wp query posts used versions wordpress version yoast seo version
| 0
|
9,424
| 12,417,445,228
|
IssuesEvent
|
2020-05-22 20:44:05
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Which tasks work out-of-the-box with environments?
|
Pri1 devops-cicd-process/tech devops/prod
|
So far the only task I've been able to find is "IIS web app deploy" (IISWebAppDeploymentOnMachineGroup). I need a task for doing file copy to the target server(s) of the environment, but the ones I've looked at all require specifying machine host name/IP address as well as admin credentials.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Which tasks work out-of-the-box with environments? - So far the only task I've been able to find is "IIS web app deploy" (IISWebAppDeploymentOnMachineGroup). I need a task for doing file copy to the target server(s) of the environment, but the ones I've looked at all require specifying machine host name/IP address as well as admin credentials.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
which tasks work out of the box with environments so far the only task i ve been able to find is iis web app deploy iiswebappdeploymentonmachinegroup i need a task for doing file copy to the target server s of the environment but the ones i ve looked at all require specifying machine host name ip address as well as admin credentials document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
2,980
| 5,965,663,282
|
IssuesEvent
|
2017-05-30 12:14:35
|
openvstorage/alba
|
https://api.github.com/repos/openvstorage/alba
|
closed
|
Failure "chunk 0 failed" must be more detailed when environment is full
|
process_wontfix type_enhancement
|
Failure "chunk 0 failed" must be more detailed when environment is full
Alba version: 1.3.6
Situation: the environment got full and the policy (12/4/14/2 - 10 nodes) could not be satisfied (disks 99% on 5 nodes). I checked the logfiles and found the below error in the alba-proxy.
```
(Failure "chunk 0 failed")
```
Detailed logview:
```
2017-02-20 10:32:43 981517 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654778 - error - Unexpected exception in proxy while handling request: (Failure "chunk 0 failed"); backtrace:; Raised at file "queue.ml", line 68, characters 17-22; Called from file "src/tools/lwt_pool2.ml", line 98, characters 25-46
2017-02-20 10:32:43 981637 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654779 - error - Request ApplySequence ("75450521-1cde-4f82-88bf-2332ffff016d",false,[(Nsm_model.Assert.ObjectHasChecksum ("owner_tag",; Sha1 c66c65175fecc3103b3b587be9b5b230889c8628)); ],[(Proxy_protocol.Protocol.Update.UploadObjectFromFile; ("00_00000049_00",; "/mnt/ssd4/vmstor_write_sco_1/75450521-1cde-4f82-88bf-2332ffff016d/00_00000049_00",; (Some Crc32c 0xdb988818))); ]) errored and took 0.189443
```
The first "chunk 0 failed" error started right after lot of the following messages:
```
2017-02-19 20:04:18 686779 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1408944 - warning - fragment upload failed:Asd_protocol.Protocol.Error.Exn(1)
```
ASD on one of the nodes that got filled up to 99%:
```
2017-02-19 20:04:22 994301 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35688 - info - returning error Asd_protocol.Protocol.Error.Full
2017-02-19 20:04:23 302203 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35690 - info - returning error Asd_protocol.Protocol.Error.Full
2017-02-19 20:04:23 848663 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35691 - info - returning error Asd_protocol.Protocol.Error.Full
```
|
1.0
|
Failure "chunk 0 failed" must be more detailed when environment is full - Failure "chunk 0 failed" must be more detailed when environment is full
Alba version: 1.3.6
Situation: the environment got full and the policy (12/4/14/2 - 10 nodes) could not be satisfied (disks 99% on 5 nodes). I checked the logfiles and found the below error in the alba-proxy.
```
(Failure "chunk 0 failed")
```
Detailed logview:
```
2017-02-20 10:32:43 981517 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654778 - error - Unexpected exception in proxy while handling request: (Failure "chunk 0 failed"); backtrace:; Raised at file "queue.ml", line 68, characters 17-22; Called from file "src/tools/lwt_pool2.ml", line 98, characters 25-46
2017-02-20 10:32:43 981637 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654779 - error - Request ApplySequence ("75450521-1cde-4f82-88bf-2332ffff016d",false,[(Nsm_model.Assert.ObjectHasChecksum ("owner_tag",; Sha1 c66c65175fecc3103b3b587be9b5b230889c8628)); ],[(Proxy_protocol.Protocol.Update.UploadObjectFromFile; ("00_00000049_00",; "/mnt/ssd4/vmstor_write_sco_1/75450521-1cde-4f82-88bf-2332ffff016d/00_00000049_00",; (Some Crc32c 0xdb988818))); ]) errored and took 0.189443
```
The first "chunk 0 failed" error started right after lot of the following messages:
```
2017-02-19 20:04:18 686779 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1408944 - warning - fragment upload failed:Asd_protocol.Protocol.Error.Exn(1)
```
ASD on one of the nodes that got filled up to 99%:
```
2017-02-19 20:04:22 994301 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35688 - info - returning error Asd_protocol.Protocol.Error.Full
2017-02-19 20:04:23 302203 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35690 - info - returning error Asd_protocol.Protocol.Error.Full
2017-02-19 20:04:23 848663 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35691 - info - returning error Asd_protocol.Protocol.Error.Full
```
|
process
|
failure chunk failed must be more detailed when environment is full failure chunk failed must be more detailed when environment is full alba version situation the environment got full and the policy nodes could not be satisfied disks on nodes i checked the logfiles and found the below error in the alba proxy failure chunk failed detailed logview stor be alba proxy error unexpected exception in proxy while handling request failure chunk failed backtrace raised at file queue ml line characters called from file src tools lwt ml line characters stor be alba proxy error request applysequence false errored and took the first chunk failed error started right after lot of the following messages stor be alba proxy warning fragment upload failed asd protocol protocol error exn asd on one of the nodes that got filled up to cpu be alba asd info returning error asd protocol protocol error full cpu be alba asd info returning error asd protocol protocol error full cpu be alba asd info returning error asd protocol protocol error full
| 1
|
7,967
| 11,147,910,893
|
IssuesEvent
|
2019-12-23 14:03:51
|
prisma/prisma2
|
https://api.github.com/repos/prisma/prisma2
|
closed
|
`prisma2 init` shortly shows `yarn install` stuff
|
bug/2-confirmed kind/bug process/candidate topic: cli-init
|

(Very possible this is a Windows only thing, ran this in `cmd.exe` - I have never seen this before preview-13 though)
`prisma2@2.0.0-preview-13.1, binary version: 14576370b55d5152d9cb90f4f509bfa56760dbbd`
|
1.0
|
`prisma2 init` shortly shows `yarn install` stuff - 
(Very possible this is a Windows only thing, ran this in `cmd.exe` - I have never seen this before preview-13 though)
`prisma2@2.0.0-preview-13.1, binary version: 14576370b55d5152d9cb90f4f509bfa56760dbbd`
|
process
|
init shortly shows yarn install stuff very possible this is a windows only thing ran this in cmd exe i have never seen this before preview though preview binary version
| 1
|
136,613
| 19,899,403,693
|
IssuesEvent
|
2022-01-25 05:28:25
|
chapel-lang/chapel
|
https://api.github.com/repos/chapel-lang/chapel
|
closed
|
Implement 'atomic' statements and transactional memory
|
type: Design area: Language
|
In Chapel's language specification, in particular section 25.10, it makes mention of `atomic` statements. It does not specify that `atomic` blocks need to use transactional memory, but it would benefit greatly from doing so. I've been experimenting with [STM](https://github.com/LouisJenkinsCS/Chapel-Atomic-Objects/blob/master/STM.chpl) in Chapel in the past, but unfortunately due to the lack of pointer-arithmetic on wide pointers (I.E pointers to remote memory) and the lack of support for atomic operations on class instances (which I have also experimented with in the past in issue #6663 and other experiments I won't mention), it isn't really possible to go further with the idea from the module level. While [others have attempted it from the lower levels](https://github.com/chapel-lang/chapel-attic/tree/collaborations/gtm), Chapel still lacks any transactional memory support.
I believe that this is something that should be done within the next few releases... I do plan on digging into it myself within the next few months after I am no longer under contract, but it would be nice to know if there is any potential discussion to be had.
|
1.0
|
Implement 'atomic' statements and transactional memory - In Chapel's language specification, in particular section 25.10, it makes mention of `atomic` statements. It does not specify that `atomic` blocks need to use transactional memory, but it would benefit greatly from doing so. I've been experimenting with [STM](https://github.com/LouisJenkinsCS/Chapel-Atomic-Objects/blob/master/STM.chpl) in Chapel in the past, but unfortunately due to the lack of pointer-arithmetic on wide pointers (I.E pointers to remote memory) and the lack of support for atomic operations on class instances (which I have also experimented with in the past in issue #6663 and other experiments I won't mention), it isn't really possible to go further with the idea from the module level. While [others have attempted it from the lower levels](https://github.com/chapel-lang/chapel-attic/tree/collaborations/gtm), Chapel still lacks any transactional memory support.
I believe that this is something that should be done within the next few releases... I do plan on digging into it myself within the next few months after I am no longer under contract, but it would be nice to know if there is any potential discussion to be had.
|
non_process
|
implement atomic statements and transactional memory in chapel s language specification in particular section it makes mention of atomic statements it does not specify that atomic blocks need to use transactional memory but it would benefit greatly from doing so i ve been experimenting with in chapel in the past but unfortunately due to the lack of pointer arithmetic on wide pointers i e pointers to remote memory and the lack of support for atomic operations on class instances which i have also experimented with in the past in issue and other experiments i won t mention it isn t really possible to go further with the idea from the module level while chapel still lacks any transactional memory support i believe that this is something that should be done within the next few releases i do plan on digging into it myself within the next few months after i am no longer under contract but it would be nice to know if there is any potential discussion to be had
| 0
|
317,231
| 27,221,149,177
|
IssuesEvent
|
2023-02-21 05:30:29
|
wpfoodmanager/wp-food-manager
|
https://api.github.com/repos/wpfoodmanager/wp-food-manager
|
closed
|
Food category and Type listing design is not proper
|
Bug In Testing
|
Food category and type listing page design is not proper. Make it proper like food listing




|
1.0
|
Food category and Type listing design is not proper - Food category and type listing page design is not proper. Make it proper like food listing




|
non_process
|
food category and type listing design is not proper food category and type listing page design is not proper make it proper like food listing
| 0
|
12,770
| 15,148,878,639
|
IssuesEvent
|
2021-02-11 11:15:07
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
White points when demosaicing with 'local average'
|
bug: pending reproduce: peculiar scope: image processing
|
**Describe the bug**
I was working on an image from PlayRaw at pixls.us. After switching demosaicing from PPG to AMaZE many big white spots appeared all over the image.
Happens when 'match greens' is set to 'local average' or 'full and local average' with demosaicing methods 'AMaZe', 'PPG' and 'VNG4'

XMP:
(https://github.com/darktable-org/darktable/files/3244592/20190531-092259-LackawannaStatePark_BullHillTrail-0037.cr2.zip)
RAW file:
https://discuss.pixls.us/t/play-raw-salamander-in-the-woods/
In exported images the white points are small but still present
**To Reproduce**
* Load RAW file with attached XMP
* Set 'match greens' in demosaicing to 'local average' or 'full and local average'
**Expected behavior**
Since the Darktable manual does not tell anything about white points that appear, these white spots should not be there
**Platform (please complete the following information):**
Darktable 2.6.2 from PPA on Ubuntu 18.04
|
1.0
|
White points when demosaicing with 'local average' - **Describe the bug**
I was working on an image from PlayRaw at pixls.us. After switching demosaicing from PPG to AMaZE many big white spots appeared all over the image.
Happens when 'match greens' is set to 'local average' or 'full and local average' with demosaicing methods 'AMaZe', 'PPG' and 'VNG4'

XMP:
(https://github.com/darktable-org/darktable/files/3244592/20190531-092259-LackawannaStatePark_BullHillTrail-0037.cr2.zip)
RAW file:
https://discuss.pixls.us/t/play-raw-salamander-in-the-woods/
In exported images the white points are small but still present
**To Reproduce**
* Load RAW file with attached XMP
* Set 'match greens' in demosaicing to 'local average' or 'full and local average'
**Expected behavior**
Since the Darktable manual does not tell anything about white points that appear, these white spots should not be there
**Platform (please complete the following information):**
Darktable 2.6.2 from PPA on Ubuntu 18.04
|
process
|
white points when demosaicing with local average describe the bug i was working on an image from playraw at pixls us after switching demosaicing from ppg to amaze many big white spots appeared all over the image happens when match greens is set to local average or full and local average with demosaicing methods amaze ppg and xmp raw file in exported images the white points are small but still present to reproduce load raw file with attached xmp set match greens in demosaicing to local average or full and local average expected behavior since the darktable manual does not tell anything about white points that appear these white spots should not be there platform please complete the following information darktable from ppa on ubuntu
| 1
|
20,569
| 27,229,965,131
|
IssuesEvent
|
2023-02-21 12:31:03
|
corona-warn-app/cwa-wishlist
|
https://api.github.com/repos/corona-warn-app/cwa-wishlist
|
closed
|
Clear statement after exposure submission that others were warned
|
enhancement mirrored-to-jira Test/Share process
|
## Current Implementation
Currently, the app only makes vague statements that others were warned after the "Warn others" process was finished successfully. Actually, the only thing I found was this message: "Thanks to your support, other people can now be warned and respond appropriately.".
## Suggested Enhancement
Make a clear statement whether others were already warned or not, like.
"The process to warn others was finished successfully and other people are now warned.".
## Expected Benefits
I wrote with quite a few users on Twitter via DM who were unsure if the app did warn others already, or if they still have to do anything. This is especially a problem if the app warned automatically in the background.
---
Internal Tracking ID: [EXPOSUREAPP-13576](https://jira-ibs.wbs.net.sap/browse/EXPOSUREAPP-13576)
|
1.0
|
Clear statement after exposure submission that others were warned - ## Current Implementation
Currently, the app only makes vague statements that others were warned after the "Warn others" process was finished successfully. Actually, the only thing I found was this message: "Thanks to your support, other people can now be warned and respond appropriately.".
## Suggested Enhancement
Make a clear statement whether others were already warned or not, like.
"The process to warn others was finished successfully and other people are now warned.".
## Expected Benefits
I wrote with quite a few users on Twitter via DM who were unsure if the app did warn others already, or if they still have to do anything. This is especially a problem if the app warned automatically in the background.
---
Internal Tracking ID: [EXPOSUREAPP-13576](https://jira-ibs.wbs.net.sap/browse/EXPOSUREAPP-13576)
|
process
|
clear statement after exposure submission that others were warned current implementation currently the app only makes vague statements that others were warned after the warn others process was finished successfully actually the only thing i found was this message thanks to your support other people can now be warned and respond appropriately suggested enhancement make a clear statement whether others were already warned or not like the process to warn others was finished successfully and other people are now warned expected benefits i wrote with quite a few users on twitter via dm who were unsure if the app did warn others already or if they still have to do anything this is especially a problem if the app warned automatically in the background internal tracking id
| 1
|
2,400
| 5,193,008,450
|
IssuesEvent
|
2017-01-22 15:08:16
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Uncaught TypeError when doing Insert Row->Click Save. v3.0.0-beta.5
|
bug inprocess
|
ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined.
This happens when I use insert row and click save. The error happens, as shown above, in ToolBar.js, line 256 when reset form is called `this.refs.form.reset();`.
ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined
at ToolBar.__handleSaveBtnClick__REACT_HOT_LOADER__ (ToolBar.js:256)
at Object.ToolBar._this.handleSaveBtnClick [as onSave] (ToolBar.js:81)
at InsertModal.__handleSave__REACT_HOT_LOADER__ (InsertModal.js:63)
at InsertModal._this.handleSave (InsertModal.js:54)
at InsertModalFooter.__handleSaveBtnClick__REACT_HOT_LOADER__ (InsertModalFooter.js:68)
at InsertModalFooter._this.handleSaveBtnClick (InsertModalFooter.js:46)
at Object.ReactErrorUtils.invokeGuardedCallback (ReactErrorUtils.js:71)
at executeDispatch (EventPluginUtils.js:79)
at Object.executeDispatchesInOrder (EventPluginUtils.js:102)
at executeDispatchesAndRelease (EventPluginHub.js:43)
at executeDispatchesAndReleaseTopLevel (EventPluginHub.js:54)
|
1.0
|
Uncaught TypeError when doing Insert Row->Click Save. v3.0.0-beta.5 - ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined.
This happens when I use insert row and click save. The error happens, as shown above, in ToolBar.js, line 256 when reset form is called `this.refs.form.reset();`.
ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined
at ToolBar.__handleSaveBtnClick__REACT_HOT_LOADER__ (ToolBar.js:256)
at Object.ToolBar._this.handleSaveBtnClick [as onSave] (ToolBar.js:81)
at InsertModal.__handleSave__REACT_HOT_LOADER__ (InsertModal.js:63)
at InsertModal._this.handleSave (InsertModal.js:54)
at InsertModalFooter.__handleSaveBtnClick__REACT_HOT_LOADER__ (InsertModalFooter.js:68)
at InsertModalFooter._this.handleSaveBtnClick (InsertModalFooter.js:46)
at Object.ReactErrorUtils.invokeGuardedCallback (ReactErrorUtils.js:71)
at executeDispatch (EventPluginUtils.js:79)
at Object.executeDispatchesInOrder (EventPluginUtils.js:102)
at executeDispatchesAndRelease (EventPluginHub.js:43)
at executeDispatchesAndReleaseTopLevel (EventPluginHub.js:54)
|
process
|
uncaught typeerror when doing insert row click save beta toolbar js uncaught typeerror cannot read property reset of undefined this happens when i use insert row and click save the error happens as shown above in toolbar js line when reset form is called this refs form reset toolbar js uncaught typeerror cannot read property reset of undefined at toolbar handlesavebtnclick react hot loader toolbar js at object toolbar this handlesavebtnclick toolbar js at insertmodal handlesave react hot loader insertmodal js at insertmodal this handlesave insertmodal js at insertmodalfooter handlesavebtnclick react hot loader insertmodalfooter js at insertmodalfooter this handlesavebtnclick insertmodalfooter js at object reacterrorutils invokeguardedcallback reacterrorutils js at executedispatch eventpluginutils js at object executedispatchesinorder eventpluginutils js at executedispatchesandrelease eventpluginhub js at executedispatchesandreleasetoplevel eventpluginhub js
| 1
|
40,240
| 16,431,631,640
|
IssuesEvent
|
2021-05-20 02:58:18
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
Support additional parameters like hbi_workspace in azurerm_machine_learning_workspace
|
enhancement service/machine-learning
|
Some additional parameters needs to be updated as per - https://docs.microsoft.com/en-us/azure/templates/microsoft.machinelearningservices/2020-03-01/workspaces
|
1.0
|
Support additional parameters like hbi_workspace in azurerm_machine_learning_workspace - Some additional parameters needs to be updated as per - https://docs.microsoft.com/en-us/azure/templates/microsoft.machinelearningservices/2020-03-01/workspaces
|
non_process
|
support additional parameters like hbi workspace in azurerm machine learning workspace some additional parameters needs to be updated as per
| 0
|
10,669
| 13,457,330,206
|
IssuesEvent
|
2020-09-09 09:03:11
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面
|
processing
|
**问题描述**
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面
**环境信息描述**
1. Window 微信客户端=》小程序面板=》最近使用小程序=》点击打开小程序。
**最简复现demo**
Window 微信客户端 打开滴滴出行小程序也是空白
|
1.0
|
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 - **问题描述**
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面
**环境信息描述**
1. Window 微信客户端=》小程序面板=》最近使用小程序=》点击打开小程序。
**最简复现demo**
Window 微信客户端 打开滴滴出行小程序也是空白
|
process
|
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 问题描述 使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 环境信息描述 window 微信客户端 》小程序面板 》最近使用小程序 》点击打开小程序。 最简复现demo window 微信客户端 打开滴滴出行小程序也是空白
| 1
|
362,967
| 25,401,239,001
|
IssuesEvent
|
2022-11-22 12:17:41
|
ooni/ooni.org
|
https://api.github.com/repos/ooni/ooni.org
|
opened
|
Documentation on how to add/update fingerprints of blocking pages
|
user feedback documentation priority/medium
|
Suggestion from the community member:
> I think we need clear documentation on how to add/update fingerprints of blocking pages. It could be prominently placed somewhere on the website in a call to action section.
The documentation can be added to the 'Contribute to the test lists' section and to the updated version of the 'Get involved' page.
|
1.0
|
Documentation on how to add/update fingerprints of blocking pages - Suggestion from the community member:
> I think we need clear documentation on how to add/update fingerprints of blocking pages. It could be prominently placed somewhere on the website in a call to action section.
The documentation can be added to the 'Contribute to the test lists' section and to the updated version of the 'Get involved' page.
|
non_process
|
documentation on how to add update fingerprints of blocking pages suggestion from the community member i think we need clear documentation on how to add update fingerprints of blocking pages it could be prominently placed somewhere on the website in a call to action section the documentation can be added to the contribute to the test lists section and to the updated version of the get involved page
| 0
|
16,428
| 21,300,224,012
|
IssuesEvent
|
2022-04-15 01:24:25
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
Consider swapping css-modules tooling for modular-css
|
:speech_balloon: RFC CSS Preprocessing ✨ Parcel 2 🔌 Parcel 2 Plugins
|
# 💬 RFC
Proposal, swap out css-modules toolchain for modular-css.
## 🔦 Context
Parcel currently has support for css-modules (yay!), but it's through very unmaintained postcss plugins from the css-modules org. These tools are pretty out of date at this point, and most of the interesting modern support for css-modules basically only exists in webpacks css-loader. Stuff, like interpolating values in selectors, (which we depend on to make component interpolation work in [astroturf](https://github.com/4Catalyzer/astroturf))
Overall the original the css-modules tooling is mostly abandoned. I've made a few attempts to try and revive/takeover/maintain these repo's but everyone if fairly unresponsive, so it seems likely they will stay that way. Since no one maintains, the pure postcss tools anymore parcel support will get more and more out of date, even if there is some life in the specific plugins that css-loader uses.
All that to say, that @tivac maintains [modular-css](https://github.com/tivac/modular-css) which is functionally the same as css-modules, but with sharp edges filed down, and active development. It's not a drop in replacement, but it is very close, actively maintained, and functionally better (both to use and consume), It'd be awesome to adapt it as the basis for css module support in Parcel.
it would improve the support for folks using css-modules as well as tooling like astroturf which is currently hobbled in parcel b/c of missing features.
Thanks!
|
1.0
|
Consider swapping css-modules tooling for modular-css - # 💬 RFC
Proposal, swap out css-modules toolchain for modular-css.
## 🔦 Context
Parcel currently has support for css-modules (yay!), but it's through very unmaintained postcss plugins from the css-modules org. These tools are pretty out of date at this point, and most of the interesting modern support for css-modules basically only exists in webpacks css-loader. Stuff, like interpolating values in selectors, (which we depend on to make component interpolation work in [astroturf](https://github.com/4Catalyzer/astroturf))
Overall the original the css-modules tooling is mostly abandoned. I've made a few attempts to try and revive/takeover/maintain these repo's but everyone if fairly unresponsive, so it seems likely they will stay that way. Since no one maintains, the pure postcss tools anymore parcel support will get more and more out of date, even if there is some life in the specific plugins that css-loader uses.
All that to say, that @tivac maintains [modular-css](https://github.com/tivac/modular-css) which is functionally the same as css-modules, but with sharp edges filed down, and active development. It's not a drop in replacement, but it is very close, actively maintained, and functionally better (both to use and consume), It'd be awesome to adapt it as the basis for css module support in Parcel.
it would improve the support for folks using css-modules as well as tooling like astroturf which is currently hobbled in parcel b/c of missing features.
Thanks!
|
process
|
consider swapping css modules tooling for modular css 💬 rfc proposal swap out css modules toolchain for modular css 🔦 context parcel currently has support for css modules yay but it s through very unmaintained postcss plugins from the css modules org these tools are pretty out of date at this point and most of the interesting modern support for css modules basically only exists in webpacks css loader stuff like interpolating values in selectors which we depend on to make component interpolation work in overall the original the css modules tooling is mostly abandoned i ve made a few attempts to try and revive takeover maintain these repo s but everyone if fairly unresponsive so it seems likely they will stay that way since no one maintains the pure postcss tools anymore parcel support will get more and more out of date even if there is some life in the specific plugins that css loader uses all that to say that tivac maintains which is functionally the same as css modules but with sharp edges filed down and active development it s not a drop in replacement but it is very close actively maintained and functionally better both to use and consume it d be awesome to adapt it as the basis for css module support in parcel it would improve the support for folks using css modules as well as tooling like astroturf which is currently hobbled in parcel b c of missing features thanks
| 1
|
254,379
| 8,073,429,849
|
IssuesEvent
|
2018-08-06 19:13:52
|
HealthCatalyst/healthcareai-r
|
https://api.github.com/repos/HealthCatalyst/healthcareai-r
|
closed
|
Limone integration
|
High Priority model interpretation new features
|
should be called after/separately from `predict` rather than being a switch to turn on during predict
|
1.0
|
Limone integration - should be called after/separately from `predict` rather than being a switch to turn on during predict
|
non_process
|
limone integration should be called after separately from predict rather than being a switch to turn on during predict
| 0
|
7,168
| 10,311,577,851
|
IssuesEvent
|
2019-08-29 17:43:01
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
PubSub: Release the first GA version
|
api: pubsub release blocking type: process
|
If no major issues are discovered during the ongoing bake-in period, the PubSub client should be released as GA.
Releasing a new version is done with the help of [releasetool](https://github.com/googleapis/releasetool). Prior to the GA release, the following must be done:
- Go through the checklist from the internal GA release doc, make sure everything is in order.
- Changes in the client lib files:
- Update PubSub Python lib README to indicate semver just like Java library has
- Update beta tag in PubSub Python lib README
- Update beta tag in Python libs root
- Set the release level to `ga` in `.repo-metadata.json`
- Change beta label in code samples in clients page in Google Cloud website
- Add to release notes (talk to @kir-titievsky)
- Update any documentation on Cloudsite
|
1.0
|
PubSub: Release the first GA version - If no major issues are discovered during the ongoing bake-in period, the PubSub client should be released as GA.
Releasing a new version is done with the help of [releasetool](https://github.com/googleapis/releasetool). Prior to the GA release, the following must be done:
- Go through the checklist from the internal GA release doc, make sure everything is in order.
- Changes in the client lib files:
- Update PubSub Python lib README to indicate semver just like Java library has
- Update beta tag in PubSub Python lib README
- Update beta tag in Python libs root
- Set the release level to `ga` in `.repo-metadata.json`
- Change beta label in code samples in clients page in Google Cloud website
- Add to release notes (talk to @kir-titievsky)
- Update any documentation on Cloudsite
|
process
|
pubsub release the first ga version if no major issues are discovered during the ongoing bake in period the pubsub client should be released as ga releasing a new version is done with the help of prior to the ga release the following must be done go through the checklist from the internal ga release doc make sure everything is in order changes in the client lib files update pubsub python lib readme to indicate semver just like java library has update beta tag in pubsub python lib readme update beta tag in python libs root set the release level to ga in repo metadata json change beta label in code samples in clients page in google cloud website add to release notes talk to kir titievsky update any documentation on cloudsite
| 1
|
337,599
| 24,547,309,918
|
IssuesEvent
|
2022-10-12 09:46:44
|
hotwax/inventory-count
|
https://api.github.com/repos/hotwax/inventory-count
|
closed
|
Readme file should be updated with build and run setup in hacktoberfest branch
|
documentation hacktoberfest
|
## What is the motivation for adding/enhancing this feature?
<!-- Describe the motivation or the concrete use case for a new feature or why one of the current ones should be enhanced. -->
Current readme file is not understandable. Make it similar as the readme file of preorder app https://github.com/hotwax/preorder of hacktoberfest branch
## What are the acceptance criteria?
<!-- List the acceptance criteria for this task in the form of a list. -->
- [ ] ...
## Can you complete this feature request by yourself?
- [ ] YES
- [ ] NO
## Additional information
<!-- If you think that any additional information would be useful, please provide them here. -->
|
1.0
|
Readme file should be updated with build and run setup in hacktoberfest branch - ## What is the motivation for adding/enhancing this feature?
<!-- Describe the motivation or the concrete use case for a new feature or why one of the current ones should be enhanced. -->
Current readme file is not understandable. Make it similar as the readme file of preorder app https://github.com/hotwax/preorder of hacktoberfest branch
## What are the acceptance criteria?
<!-- List the acceptance criteria for this task in the form of a list. -->
- [ ] ...
## Can you complete this feature request by yourself?
- [ ] YES
- [ ] NO
## Additional information
<!-- If you think that any additional information would be useful, please provide them here. -->
|
non_process
|
readme file should be updated with build and run setup in hacktoberfest branch what is the motivation for adding enhancing this feature current readme file is not understandable make it similar as the readme file of preorder app of hacktoberfest branch what are the acceptance criteria can you complete this feature request by yourself yes no additional information
| 0
|
16,762
| 21,934,483,325
|
IssuesEvent
|
2022-05-23 12:43:23
|
camunda/zeebe-process-test
|
https://api.github.com/repos/camunda/zeebe-process-test
|
opened
|
Test case with multiple workers is flaky
|
kind/bug team/process-automation
|
I have a strange problem in https://github.com/camunda-community-hub/spring-zeebe/ with testing where I could need some help. I currently guess it is rooted in zeebe-process-test, but are not sure.
In some wired situations (which I currently haven't tracked down exactly) a job worker stops polling according to the trace log of zeebe-process-test. I can't see anything on the worker side - and I wondered if in EZE a JobAcquisition could hang for file (for whatever reason)?

it happens when I run the test in https://github.com/berndruecker/customer-onboarding-camunda-8-springboot-extended/blob/master/src/test/java/io/berndruecker/onboarding/customer/TestCustomerOnboardingProcess.java - but not every time...
|
1.0
|
Test case with multiple workers is flaky - I have a strange problem in https://github.com/camunda-community-hub/spring-zeebe/ with testing where I could need some help. I currently guess it is rooted in zeebe-process-test, but are not sure.
In some wired situations (which I currently haven't tracked down exactly) a job worker stops polling according to the trace log of zeebe-process-test. I can't see anything on the worker side - and I wondered if in EZE a JobAcquisition could hang for file (for whatever reason)?

it happens when I run the test in https://github.com/berndruecker/customer-onboarding-camunda-8-springboot-extended/blob/master/src/test/java/io/berndruecker/onboarding/customer/TestCustomerOnboardingProcess.java - but not every time...
|
process
|
test case with multiple workers is flaky i have a strange problem in with testing where i could need some help i currently guess it is rooted in zeebe process test but are not sure in some wired situations which i currently haven t tracked down exactly a job worker stops polling according to the trace log of zeebe process test i can t see anything on the worker side and i wondered if in eze a jobacquisition could hang for file for whatever reason it happens when i run the test in but not every time
| 1
|
350
| 2,793,381,240
|
IssuesEvent
|
2015-05-11 10:34:54
|
ecodistrict/IDSSDashboard
|
https://api.github.com/repos/ecodistrict/IDSSDashboard
|
closed
|
Show what data is required by module
|
enhancement form feedback 09102014 process step: collect data
|
It would be good if each module would show what data it requires and what data is used by default, and also what data has been uploaded which it is using.
|
1.0
|
Show what data is required by module - It would be good if each module would show what data it requires and what data is used by default, and also what data has been uploaded which it is using.
|
process
|
show what data is required by module it would be good if each module would show what data it requires and what data is used by default and also what data has been uploaded which it is using
| 1
|
12,646
| 15,019,967,845
|
IssuesEvent
|
2021-02-01 14:11:02
|
Talendar/nevopy
|
https://api.github.com/repos/Talendar/nevopy
|
opened
|
Add a network processing scheduler
|
comp:processing type:feature
|
**Feature:**
A processing scheduler (subclass of [`ProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/base_scheduler.py#L47)) capable of assigning work to different machines on a network. The [`RayProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/ray_processing.py#L42) seems to be able to do that to some extent, but, although I could make it work on a local network, I was unable to use it to split the processing among computers in the Internet.
The network processing scheduler must be able to manage other processing schedulers (workers), each of which is in a different machine.
|
1.0
|
Add a network processing scheduler - **Feature:**
A processing scheduler (subclass of [`ProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/base_scheduler.py#L47)) capable of assigning work to different machines on a network. The [`RayProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/ray_processing.py#L42) seems to be able to do that to some extent, but, although I could make it work on a local network, I was unable to use it to split the processing among computers in the Internet.
The network processing scheduler must be able to manage other processing schedulers (workers), each of which is in a different machine.
|
process
|
add a network processing scheduler feature a processing scheduler subclass of capable of assigning work to different machines on a network the seems to be able to do that to some extent but although i could make it work on a local network i was unable to use it to split the processing among computers in the internet the network processing scheduler must be able to manage other processing schedulers workers each of which is in a different machine
| 1
|
845
| 3,314,805,819
|
IssuesEvent
|
2015-11-06 08:19:07
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
opened
|
Implement "select USER()"
|
MYSQL PROTOCOL QUERY PROCESSOR ROUTING
|
## WHY
Some clients issue a "select USER()" . This could be executed directly from ProxySQL
## WHAT
* [ ] extend MySQL_Session::handler_special_queries() to also support "select USER()"
|
1.0
|
Implement "select USER()" - ## WHY
Some clients issue a "select USER()" . This could be executed directly from ProxySQL
## WHAT
* [ ] extend MySQL_Session::handler_special_queries() to also support "select USER()"
|
process
|
implement select user why some clients issue a select user this could be executed directly from proxysql what extend mysql session handler special queries to also support select user
| 1
|
3,852
| 6,808,552,216
|
IssuesEvent
|
2017-11-04 04:30:06
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
reopened
|
Some command lines not working correctly
|
apps-makeClass status-inprocess type-bug
|
Some of the older tools/apps such as `makeClass`, `grabABI`, and `ethslurp` may accept -commands with a single dash instead of --command (double dash). Check all the options.cpp files for this error.
|
1.0
|
Some command lines not working correctly - Some of the older tools/apps such as `makeClass`, `grabABI`, and `ethslurp` may accept -commands with a single dash instead of --command (double dash). Check all the options.cpp files for this error.
|
process
|
some command lines not working correctly some of the older tools apps such as makeclass grababi and ethslurp may accept commands with a single dash instead of command double dash check all the options cpp files for this error
| 1
|
8,763
| 11,883,162,081
|
IssuesEvent
|
2020-03-27 15:31:06
|
prisma/prisma-client-js
|
https://api.github.com/repos/prisma/prisma-client-js
|
closed
|
Results should always include foreign key columns
|
kind/feature process/product
|
I would like to propose that when returning records from a table, any foreign key columns are automatically included. For example:
```
model User {
id Int @id @unique
name String
posts Post[]
}
model Post {
id Int @id @unique
text String
user User
}
```
Returning all posts with `photon.posts.findMany()` today results in:
```json
[
{
"id": 1,
"text": "Hello, world"
},
{
"id": 2,
"text": "Foobar"
}
]
```
I'm proposing that it return the following instead:
```json
[
{
"id": 1,
"text": "Hello, world",
"user": 10
},
{
"id": 2,
"text": "Foobar",
"user": 11
}
]
```
I can still return the full associated User object, if desired, with the `include:` option.
My reasoning is that when I say "I want a post" or "I want all posts" then that means I want *all* of the data for those posts. If I only want a subset of columns then I'll use the `select:` option and list the specific fields that I want.
In the app I'm currently working in I'll pull a collection of data when the app launches and additional data later if the user navigates to a certain page. When they do, the additional data I need is looked up by a foreign key and then associated and displayed. I find myself constantly having to use an `:include => :select` in order to get that foreign key returned in my first batch of data. One of my queries looks like:
```javascript
photon.users.findMany({
include: {
note: { select: { id: true } },
responses: {
include: {
form: { select: { id: true } },
answers: {
include: {
question: { select: { id: true } }
}
}
}
}
}
})
```
If the foreign keys were included by default this would become:
```javascript
photon.users.findMany({
include: {
note: true,
responses: { include: { answers: true } }
}
})
```
|
1.0
|
Results should always include foreign key columns - I would like to propose that when returning records from a table, any foreign key columns are automatically included. For example:
```
model User {
id Int @id @unique
name String
posts Post[]
}
model Post {
id Int @id @unique
text String
user User
}
```
Returning all posts with `photon.posts.findMany()` today results in:
```json
[
{
"id": 1,
"text": "Hello, world"
},
{
"id": 2,
"text": "Foobar"
}
]
```
I'm proposing that it return the following instead:
```json
[
{
"id": 1,
"text": "Hello, world",
"user": 10
},
{
"id": 2,
"text": "Foobar",
"user": 11
}
]
```
I can still return the full associated User object, if desired, with the `include:` option.
My reasoning is that when I say "I want a post" or "I want all posts" then that means I want *all* of the data for those posts. If I only want a subset of columns then I'll use the `select:` option and list the specific fields that I want.
In the app I'm currently working in I'll pull a collection of data when the app launches and additional data later if the user navigates to a certain page. When they do, the additional data I need is looked up by a foreign key and then associated and displayed. I find myself constantly having to use an `:include => :select` in order to get that foreign key returned in my first batch of data. One of my queries looks like:
```javascript
photon.users.findMany({
include: {
note: { select: { id: true } },
responses: {
include: {
form: { select: { id: true } },
answers: {
include: {
question: { select: { id: true } }
}
}
}
}
}
})
```
If the foreign keys were included by default this would become:
```javascript
photon.users.findMany({
include: {
note: true,
responses: { include: { answers: true } }
}
})
```
|
process
|
results should always include foreign key columns i would like to propose that when returning records from a table any foreign key columns are automatically included for example model user id int id unique name string posts post model post id int id unique text string user user returning all posts with photon posts findmany today results in json id text hello world id text foobar i m proposing that it return the following instead json id text hello world user id text foobar user i can still return the full associated user object if desired with the include option my reasoning is that when i say i want a post or i want all posts then that means i want all of the data for those posts if i only want a subset of columns then i ll use the select option and list the specific fields that i want in the app i m currently working in i ll pull a collection of data when the app launches and additional data later if the user navigates to a certain page when they do the additional data i need is looked up by a foreign key and then associated and displayed i find myself constantly having to use an include select in order to get that foreign key returned in my first batch of data one of my queries looks like javascript photon users findmany include note select id true responses include form select id true answers include question select id true if the foreign keys were included by default this would become javascript photon users findmany include note true responses include answers true
| 1
|
4,372
| 7,260,515,864
|
IssuesEvent
|
2018-02-18 10:54:36
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE][processing] New algorithm for subdividing geometries
|
Automatic new feature Processing User Manual
|
Original commit: https://github.com/qgis/QGIS/commit/4dcf8d8d10249971f5ec1358a172078e11e6bad8 by nyalldawson
The returned geometry will be a collection containing subdivided parts
from the original geometry, where no part has more then the specified
maximum number of nodes.
This is useful for dividing a complex geometry into less complex parts,
which are better able to be spatially indexed and faster to perform
further operations such as intersects on. The returned geometry parts may
not be valid and may contain self-intersections.
|
1.0
|
[FEATURE][processing] New algorithm for subdividing geometries - Original commit: https://github.com/qgis/QGIS/commit/4dcf8d8d10249971f5ec1358a172078e11e6bad8 by nyalldawson
The returned geometry will be a collection containing subdivided parts
from the original geometry, where no part has more then the specified
maximum number of nodes.
This is useful for dividing a complex geometry into less complex parts,
which are better able to be spatially indexed and faster to perform
further operations such as intersects on. The returned geometry parts may
not be valid and may contain self-intersections.
|
process
|
new algorithm for subdividing geometries original commit by nyalldawson the returned geometry will be a collection containing subdivided parts from the original geometry where no part has more then the specified maximum number of nodes this is useful for dividing a complex geometry into less complex parts which are better able to be spatially indexed and faster to perform further operations such as intersects on the returned geometry parts may not be valid and may contain self intersections
| 1
|
80,669
| 23,276,156,690
|
IssuesEvent
|
2022-08-05 07:24:48
|
reitmas32/Next
|
https://api.github.com/repos/reitmas32/Next
|
opened
|
Create a basic builder
|
builder
|
## Builder that uses nothing in the base
### Example of config.yaml
```yaml
basic_release:
base: basic
c_compiler: gcc
cxx_compiler: g++
linker: ld
files_cxx:
- main.cpp
- src/func/suma.cpp
- src/structs/*.cc
files_c:
- main_of_c.c
- src/func/suma.c
- src/structs/*.c
c_compiler_regex:
$C $FILE -o #FILE.o
cxx_compiler_regex:
$CXX $FILE -o #FILE.o
ld_regex:
$LD $FILES -lgl -pthread
```
|
1.0
|
Create a basic builder - ## Builder that uses nothing in the base
### Example of config.yaml
```yaml
basic_release:
base: basic
c_compiler: gcc
cxx_compiler: g++
linker: ld
files_cxx:
- main.cpp
- src/func/suma.cpp
- src/structs/*.cc
files_c:
- main_of_c.c
- src/func/suma.c
- src/structs/*.c
c_compiler_regex:
$C $FILE -o #FILE.o
cxx_compiler_regex:
$CXX $FILE -o #FILE.o
ld_regex:
$LD $FILES -lgl -pthread
```
|
non_process
|
create a basic builder builder that uses nothing in the base example of config yaml yaml basic release base basic c compiler gcc cxx compiler g linker ld files cxx main cpp src func suma cpp src structs cc files c main of c c src func suma c src structs c c compiler regex c file o file o cxx compiler regex cxx file o file o ld regex ld files lgl pthread
| 0
|
18,109
| 24,135,657,568
|
IssuesEvent
|
2022-09-21 11:03:50
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
ProcessPoolExecutor: float32 result of executors seems to get truncated.
|
type-bug pending expert-multiprocessing
|
# Bug report
```
import time
import numpy as np
from concurrent.futures import ProcessPoolExecutor
def example_executor():
time.sleep(1)
return np.float32(103167328256.2)
if __name__ == '__main__':
processing_pool = ProcessPoolExecutor()
future = processing_pool.submit(example_executor)
while not future.done():
time.sleep(1)
print(future.result())
```
yields:
`
103167330000.0
`
# Your environment
Windows 10. Python 3.10.2 and also 3.11.0rc2
|
1.0
|
ProcessPoolExecutor: float32 result of executors seems to get truncated. - # Bug report
```
import time
import numpy as np
from concurrent.futures import ProcessPoolExecutor
def example_executor():
time.sleep(1)
return np.float32(103167328256.2)
if __name__ == '__main__':
processing_pool = ProcessPoolExecutor()
future = processing_pool.submit(example_executor)
while not future.done():
time.sleep(1)
print(future.result())
```
yields:
`
103167330000.0
`
# Your environment
Windows 10. Python 3.10.2 and also 3.11.0rc2
|
process
|
processpoolexecutor result of executors seems to get truncated bug report import time import numpy as np from concurrent futures import processpoolexecutor def example executor time sleep return np if name main processing pool processpoolexecutor future processing pool submit example executor while not future done time sleep print future result yields your environment windows python and also
| 1
|
11,319
| 14,138,102,894
|
IssuesEvent
|
2020-11-10 07:53:16
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
QQ小程序和微信小程序的订阅消息接口差异
|
processing
|
# 描述
在 QQ小程序 中订阅消息的接口为 `qq.subscribeAppMsg` [文档](https://q.qq.com/wiki/develop/miniprogram/API/open_port/port_subscription.html)
在微信小程序中订阅消息调用为 `wx.requestSubscribeMessage` [文档](https://developers.weixin.qq.com/miniprogram/dev/api/open-api/subscribe-message/wx.requestSubscribeMessage.html)
在使用 mpx 推荐的兼容写法 `mpx.requestSubscribeMessage` 无法实现自动代理到 QQ小程序 对应的API中。对于这种情况下 mpx 目前的方案可以通过配置实现吗?
# 备注
可以在 QQ小程序 通过 `mpx.subscribeAppMsg` 调用到接,但是这样就需要做平台判断。另外,由于项目使用了 TypeScript ,这样使用会在类型推导上产生错误。
|
1.0
|
QQ小程序和微信小程序的订阅消息接口差异 - # 描述
在 QQ小程序 中订阅消息的接口为 `qq.subscribeAppMsg` [文档](https://q.qq.com/wiki/develop/miniprogram/API/open_port/port_subscription.html)
在微信小程序中订阅消息调用为 `wx.requestSubscribeMessage` [文档](https://developers.weixin.qq.com/miniprogram/dev/api/open-api/subscribe-message/wx.requestSubscribeMessage.html)
在使用 mpx 推荐的兼容写法 `mpx.requestSubscribeMessage` 无法实现自动代理到 QQ小程序 对应的API中。对于这种情况下 mpx 目前的方案可以通过配置实现吗?
# 备注
可以在 QQ小程序 通过 `mpx.subscribeAppMsg` 调用到接,但是这样就需要做平台判断。另外,由于项目使用了 TypeScript ,这样使用会在类型推导上产生错误。
|
process
|
qq小程序和微信小程序的订阅消息接口差异 描述 在 qq小程序 中订阅消息的接口为 qq subscribeappmsg 在微信小程序中订阅消息调用为 wx requestsubscribemessage 在使用 mpx 推荐的兼容写法 mpx requestsubscribemessage 无法实现自动代理到 qq小程序 对应的api中。对于这种情况下 mpx 目前的方案可以通过配置实现吗? 备注 可以在 qq小程序 通过 mpx subscribeappmsg 调用到接,但是这样就需要做平台判断。另外,由于项目使用了 typescript ,这样使用会在类型推导上产生错误。
| 1
|
10,190
| 13,045,131,038
|
IssuesEvent
|
2020-07-29 06:35:01
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
The Start/Stop VMs solution during off-hours: bug with Resource Group naming
|
Pri2 automation/svc cxp process-automation/subsvc product-question triaged
|
[Enter feedback here]
It seems that the automation is bugging with resource group with a . in the name Error message
XXX.Production is not a valid ResourceGroup Name. Please verify your input!
Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Azure Automation Start/Stop VMs during off-hours overview](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
The Start/Stop VMs solution during off-hours: bug with Resource Group naming -
[Enter feedback here]
It seems that the automation is bugging with resource group with a . in the name Error message
XXX.Production is not a valid ResourceGroup Name. Please verify your input!
Thanks
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Azure Automation Start/Stop VMs during off-hours overview](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
the start stop vms solution during off hours bug with resource group naming it seems that the automation is bugging with resource group with a in the name error message xxx production is not a valid resourcegroup name please verify your input thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
157,566
| 13,696,986,465
|
IssuesEvent
|
2020-10-01 01:40:45
|
gurupratap-matharu/volu
|
https://api.github.com/repos/gurupratap-matharu/volu
|
opened
|
Inherit base template to base_with_header and base_with_main html templates
|
documentation enhancement
|
- Use _base.html to create base_with_header html template which has a nice page header
- Similarly, use _base.html to crate a base_with_main html template which has access to directy main content in child templates
This way we do not have to maintain margin and paddings in css in all templates
Use simple template inheritance and our view templates should directly write content in specified containers without being bothered
about responsiveness and alignments
|
1.0
|
Inherit base template to base_with_header and base_with_main html templates - - Use _base.html to create base_with_header html template which has a nice page header
- Similarly, use _base.html to crate a base_with_main html template which has access to directy main content in child templates
This way we do not have to maintain margin and paddings in css in all templates
Use simple template inheritance and our view templates should directly write content in specified containers without being bothered
about responsiveness and alignments
|
non_process
|
inherit base template to base with header and base with main html templates use base html to create base with header html template which has a nice page header similarly use base html to crate a base with main html template which has access to directy main content in child templates this way we do not have to maintain margin and paddings in css in all templates use simple template inheritance and our view templates should directly write content in specified containers without being bothered about responsiveness and alignments
| 0
|
163,671
| 25,856,163,682
|
IssuesEvent
|
2022-12-13 13:55:36
|
decidim/decidim
|
https://api.github.com/repos/decidim/decidim
|
opened
|
HERE Maps static map size
|
contract: redesign
|
If you use HERE maps, currently, a small static map is generated:
<img width="811" alt="CleanShot 2022-12-13 at 14 49 54@2x" src="https://user-images.githubusercontent.com/24740/207345569-b9b47136-0a3a-41e4-9d13-34dd201037cb.png">
(https://www.decidim.barcelona/assemblies/barribesosmaresme/f/3229/meetings/5765)
In the new design, this static map is slightly bigger but as we are using the same call, the map is pixelated:
<img width="1001" alt="CleanShot 2022-12-13 at 14 54 07@2x" src="https://user-images.githubusercontent.com/24740/207347884-283fcf07-5018-4843-bfd5-bf467913b7a1.png">
The call to the map should be changed (or made dynamic?) to adapt to the new size.
https://github.com/decidim/decidim/blob/db91cf387557bfccee917e744303b25d06c691c4/decidim-generators/lib/decidim/generators/app_templates/initializer.rb#L49
https://developer.here.com/documentation/map-image/dev_guide/topics/display-options.html
|
1.0
|
HERE Maps static map size - If you use HERE maps, currently, a small static map is generated:
<img width="811" alt="CleanShot 2022-12-13 at 14 49 54@2x" src="https://user-images.githubusercontent.com/24740/207345569-b9b47136-0a3a-41e4-9d13-34dd201037cb.png">
(https://www.decidim.barcelona/assemblies/barribesosmaresme/f/3229/meetings/5765)
In the new design, this static map is slightly bigger but as we are using the same call, the map is pixelated:
<img width="1001" alt="CleanShot 2022-12-13 at 14 54 07@2x" src="https://user-images.githubusercontent.com/24740/207347884-283fcf07-5018-4843-bfd5-bf467913b7a1.png">
The call to the map should be changed (or made dynamic?) to adapt to the new size.
https://github.com/decidim/decidim/blob/db91cf387557bfccee917e744303b25d06c691c4/decidim-generators/lib/decidim/generators/app_templates/initializer.rb#L49
https://developer.here.com/documentation/map-image/dev_guide/topics/display-options.html
|
non_process
|
here maps static map size if you use here maps currently a small static map is generated img width alt cleanshot at src in the new design this static map is slightly bigger but as we are using the same call the map is pixelated img width alt cleanshot at src the call to the map should be changed or made dynamic to adapt to the new size
| 0
|
21,393
| 29,202,232,295
|
IssuesEvent
|
2023-05-21 00:37:44
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] Data Scientist na Coodesh
|
SALVADOR BANCO DE DADOS BIG DATA DATA SCIENCE SQL AWS REQUISITOS REMOTO PROCESSOS INOVAÇÃO GITHUB CI UMA QUALIDADE R BI MINERAÇÃO DE DADOS MACHINE LEARNING NEGÓCIOS IOT Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>First Decision</strong> está em busca de <strong><ins>Data Scientist</ins></strong><strong> </strong>para compor seu time!</p>
<p>A First Decision é uma empresa de tecnologia com expertise em BI, Data Governance e Data Quality. Além disso, oferecemos soluções em tecnologias como IoT, Machine Learning, Big Data, Biometria, entre outras. Somos parceiros Hitachi Vantara, AWS e Mongo DB. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos parceiros auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p>
<p></p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Analisar, modelar, propor, implementar e avaliar soluções para problemas complexos e simples de negócios, por meio de técnicas estatísticas, matemáticas e computacionais aplicadas ao processamento e visualização de dados.</li>
</ul>
## First Decision:
<p>Trazemos vantagem competitiva e auxiliamos nas tomadas de decisões para que sejam as mais inteligentes e vantajosas possíveis. Composta de colaboradores altamente capacitados e especializados, estamos sempre pesquisando e idealizando novas soluções que permitem aos nossos clientes gerenciar os dados de suas organizações para impulsionar a eficiência operacional. Através da criação e implementação de soluções que aprimorem a maneira de relacionar o seu negócio com a sociedade. Conduzimos os processos de transformação através da inovação para potencializar seus resultados. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos clientes auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p>
</p>
## Habilidades:
- Big Data Architects
- Banco de dados relacionais (SQL)
- Machine Learning
## Local:
100% Remoto
## Requisitos:
- Experiência sólida com Big Data;
- Ampla experiência em Ciência de Dados;
- Superior completo em TI;
- Domínio da linguagem SQL;
- Domínio de SGBDs;
- Conhecimento avançado em mineração de dados, extração de informações não triviais e aprendizagem de máquina.
## Benefícios:
- Vale alimentação: R$ 32,17 por dia;
- Plano de Saúde: 50% custeada pela empresa;
- Plano Odontológico;
- Plataforma para terapia;
- Seguro de vida.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Data Scientist na First Decision](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Categoria
Data Science
|
1.0
|
[Remoto] Data Scientist na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>First Decision</strong> está em busca de <strong><ins>Data Scientist</ins></strong><strong> </strong>para compor seu time!</p>
<p>A First Decision é uma empresa de tecnologia com expertise em BI, Data Governance e Data Quality. Além disso, oferecemos soluções em tecnologias como IoT, Machine Learning, Big Data, Biometria, entre outras. Somos parceiros Hitachi Vantara, AWS e Mongo DB. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos parceiros auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p>
<p></p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Analisar, modelar, propor, implementar e avaliar soluções para problemas complexos e simples de negócios, por meio de técnicas estatísticas, matemáticas e computacionais aplicadas ao processamento e visualização de dados.</li>
</ul>
## First Decision:
<p>Trazemos vantagem competitiva e auxiliamos nas tomadas de decisões para que sejam as mais inteligentes e vantajosas possíveis. Composta de colaboradores altamente capacitados e especializados, estamos sempre pesquisando e idealizando novas soluções que permitem aos nossos clientes gerenciar os dados de suas organizações para impulsionar a eficiência operacional. Através da criação e implementação de soluções que aprimorem a maneira de relacionar o seu negócio com a sociedade. Conduzimos os processos de transformação através da inovação para potencializar seus resultados. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos clientes auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p>
</p>
## Habilidades:
- Big Data Architects
- Banco de dados relacionais (SQL)
- Machine Learning
## Local:
100% Remoto
## Requisitos:
- Experiência sólida com Big Data;
- Ampla experiência em Ciência de Dados;
- Superior completo em TI;
- Domínio da linguagem SQL;
- Domínio de SGBDs;
- Conhecimento avançado em mineração de dados, extração de informações não triviais e aprendizagem de máquina.
## Benefícios:
- Vale alimentação: R$ 32,17 por dia;
- Plano de Saúde: 50% custeada pela empresa;
- Plano Odontológico;
- Plataforma para terapia;
- Seguro de vida.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Data Scientist na First Decision](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Categoria
Data Science
|
process
|
data scientist na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a first decision está em busca de data scientist para compor seu time a first decision é uma empresa de tecnologia com expertise em bi data governance e data quality além disso oferecemos soluções em tecnologias como iot machine learning big data biometria entre outras somos parceiros hitachi vantara aws e mongo db tudo isso faz da first decision uma empresa moderna e à frente do seu tempo que oferece soluções inteligentes condiciona soluções integradas e agrega valores à gestão de negócios de nossos parceiros auxiliando na sua competitividade sempre visando o desenvolvimento sustentável inovação e qualidade pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor responsabilidades analisar modelar propor implementar e avaliar soluções para problemas complexos e simples de negócios por meio de técnicas estatísticas matemáticas e computacionais aplicadas ao processamento e visualização de dados first decision trazemos vantagem competitiva e auxiliamos nas tomadas de decisões para que sejam as mais inteligentes e vantajosas possíveis composta de colaboradores altamente capacitados e especializados estamos sempre pesquisando e idealizando novas soluções que permitem aos nossos clientes gerenciar os dados de suas organizações para impulsionar a eficiência operacional através da criação e implementação de soluções que aprimorem a maneira de relacionar o seu negócio com a sociedade conduzimos os processos de transformação através da inovação para potencializar seus resultados tudo isso faz da first decision uma empresa moderna e à frente do seu tempo que oferece soluções inteligentes condiciona soluções integradas e agrega valores à gestão de negócios de nossos clientes auxiliando na sua competitividade sempre visando o desenvolvimento sustentável inovação e qualidade pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor habilidades big data architects banco de dados relacionais sql machine learning local remoto requisitos experiência sólida com big data ampla experiência em ciência de dados superior completo em ti domínio da linguagem sql domínio de sgbds conhecimento avançado em mineração de dados extração de informações não triviais e aprendizagem de máquina benefícios vale alimentação r por dia plano de saúde custeada pela empresa plano odontológico plataforma para terapia seguro de vida como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime clt categoria data science
| 1
|
333,941
| 24,397,618,737
|
IssuesEvent
|
2022-10-04 20:51:20
|
Leafwing-Studios/leafwing_input_playback
|
https://api.github.com/repos/Leafwing-Studios/leafwing_input_playback
|
closed
|
Add basic instructions to README
|
documentation
|
We should explain a bit more about how this crate is intended to do, how it works and so on.
|
1.0
|
Add basic instructions to README - We should explain a bit more about how this crate is intended to do, how it works and so on.
|
non_process
|
add basic instructions to readme we should explain a bit more about how this crate is intended to do how it works and so on
| 0
|
112,295
| 17,089,398,484
|
IssuesEvent
|
2021-07-08 15:30:20
|
keep-network/coverage-pools
|
https://api.github.com/repos/keep-network/coverage-pools
|
closed
|
Flashloan can be used to increase rewards when calling the RiskManager
|
:eyeglasses: security-audit
|
```
Severity: Medium Difficulty: Medium
Type: Data Validation
Target: RiskManager.sol, CoveragePool.sol
```
The Risk Manager rewards users that report liquidations, but they can manipulate the value received using flashloans.
`covAmountToGrant` determines the reward to grant the uses that notify when deposits are in liquidations or liquidated using `RiskManager`:
https://github.com/keep-network/coverage-pools/blob/main/contracts/RiskManagerV1.sol#L909-L943
The computation of the reward depends directly on the total supply of `underwriterToken`:
https://github.com/keep-network/coverage-pools/blob/main/contracts/CoveragePool.sol#L231-L241
A user can temporarily increase this value by making a series of calls to do the following in a single transaction:
1. Take out a flash loan
2. Make a large deposit in the underwriter token
3. Notify a liquidation and receive the reward
4. Sell the underwriter tokens in a secondary market
5. Repay the flash loan
**Exploit Scenario**
Eve wants to profit from the RiskManager so she takes a flashloan to increase her reward when reporting liquidations and is able to manipulate the system in her favor.
**Recommendations**
Short term, revise the reward computation to avoid using `assetPool.underwriterToken().totalSupply()`, which depends on the total supply of the underwriter token.
Long term, review the specification on incentives and penalties to make sure that they do not depend on variables that can be manipulated by users. Use Echidna or Manticore to test them.
|
True
|
Flashloan can be used to increase rewards when calling the RiskManager - ```
Severity: Medium Difficulty: Medium
Type: Data Validation
Target: RiskManager.sol, CoveragePool.sol
```
The Risk Manager rewards users that report liquidations, but they can manipulate the value received using flashloans.
`covAmountToGrant` determines the reward to grant the uses that notify when deposits are in liquidations or liquidated using `RiskManager`:
https://github.com/keep-network/coverage-pools/blob/main/contracts/RiskManagerV1.sol#L909-L943
The computation of the reward depends directly on the total supply of `underwriterToken`:
https://github.com/keep-network/coverage-pools/blob/main/contracts/CoveragePool.sol#L231-L241
A user can temporarily increase this value by making a series of calls to do the following in a single transaction:
1. Take out a flash loan
2. Make a large deposit in the underwriter token
3. Notify a liquidation and receive the reward
4. Sell the underwriter tokens in a secondary market
5. Repay the flash loan
**Exploit Scenario**
Eve wants to profit from the RiskManager so she takes a flashloan to increase her reward when reporting liquidations and is able to manipulate the system in her favor.
**Recommendations**
Short term, revise the reward computation to avoid using `assetPool.underwriterToken().totalSupply()`, which depends on the total supply of the underwriter token.
Long term, review the specification on incentives and penalties to make sure that they do not depend on variables that can be manipulated by users. Use Echidna or Manticore to test them.
|
non_process
|
flashloan can be used to increase rewards when calling the riskmanager severity medium difficulty medium type data validation target riskmanager sol coveragepool sol the risk manager rewards users that report liquidations but they can manipulate the value received using flashloans covamounttogrant determines the reward to grant the uses that notify when deposits are in liquidations or liquidated using riskmanager the computation of the reward depends directly on the total supply of underwritertoken a user can temporarily increase this value by making a series of calls to do the following in a single transaction take out a flash loan make a large deposit in the underwriter token notify a liquidation and receive the reward sell the underwriter tokens in a secondary market repay the flash loan exploit scenario eve wants to profit from the riskmanager so she takes a flashloan to increase her reward when reporting liquidations and is able to manipulate the system in her favor recommendations short term revise the reward computation to avoid using assetpool underwritertoken totalsupply which depends on the total supply of the underwriter token long term review the specification on incentives and penalties to make sure that they do not depend on variables that can be manipulated by users use echidna or manticore to test them
| 0
|
186,765
| 21,969,349,641
|
IssuesEvent
|
2022-05-25 01:13:30
|
DarshiL7/Authentication-demo-Project
|
https://api.github.com/repos/DarshiL7/Authentication-demo-Project
|
opened
|
Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll: 1 vulnerabilities (highest severity is: 5.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p></summary>
<p>Microsoft.AspNetCore.Authentication.JwtBearer</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p>
<p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-34532](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll | Direct | Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-34532</summary>
### Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p>
<p>Microsoft.AspNetCore.Authentication.JwtBearer</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p>
<p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p>
<p>
Dependency Hierarchy:
- :x: **Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
ASP.NET Core and Visual Studio Information Disclosure Vulnerability
<p>Publish Date: 2021-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532>CVE-2021-34532</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q7cg-43mg-qp69">https://github.com/advisories/GHSA-q7cg-43mg-qp69</a></p>
<p>Release Date: 2021-08-12</p>
<p>Fix Resolution: Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"Microsoft.AspNetCore.Authentication.JwtBearer","packageVersion":"3.100.1020.52004","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"Microsoft.AspNetCore.Authentication.JwtBearer:3.100.1020.52004","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-34532","vulnerabilityDetails":"ASP.NET Core and Visual Studio Information Disclosure Vulnerability","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}]</REMEDIATE> -->
|
True
|
Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll: 1 vulnerabilities (highest severity is: 5.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p></summary>
<p>Microsoft.AspNetCore.Authentication.JwtBearer</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p>
<p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-34532](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll | Direct | Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-34532</summary>
### Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p>
<p>Microsoft.AspNetCore.Authentication.JwtBearer</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p>
<p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p>
<p>
Dependency Hierarchy:
- :x: **Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
ASP.NET Core and Visual Studio Information Disclosure Vulnerability
<p>Publish Date: 2021-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532>CVE-2021-34532</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q7cg-43mg-qp69">https://github.com/advisories/GHSA-q7cg-43mg-qp69</a></p>
<p>Release Date: 2021-08-12</p>
<p>Fix Resolution: Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"Microsoft.AspNetCore.Authentication.JwtBearer","packageVersion":"3.100.1020.52004","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"Microsoft.AspNetCore.Authentication.JwtBearer:3.100.1020.52004","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-34532","vulnerabilityDetails":"ASP.NET Core and Visual Studio Information Disclosure Vulnerability","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}]</REMEDIATE> -->
|
non_process
|
microsoft aspnetcore authentication jwtbearer dll vulnerabilities highest severity is vulnerable library microsoft aspnetcore authentication jwtbearer dll microsoft aspnetcore authentication jwtbearer library home page a href path to vulnerable library authentication demo project api bin debug microsoft aspnetcore authentication jwtbearer dll found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium microsoft aspnetcore authentication jwtbearer dll direct microsoft aspnetcore authentication jwtbearer details cve vulnerable library microsoft aspnetcore authentication jwtbearer dll microsoft aspnetcore authentication jwtbearer library home page a href path to vulnerable library authentication demo project api bin debug microsoft aspnetcore authentication jwtbearer dll dependency hierarchy x microsoft aspnetcore authentication jwtbearer dll vulnerable library found in head commit a href found in base branch main vulnerability details asp net core and visual studio information disclosure vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore authentication jwtbearer step up your open source security game with whitesource istransitivedependency false dependencytree microsoft aspnetcore authentication jwtbearer isminimumfixversionavailable true minimumfixversion microsoft aspnetcore authentication jwtbearer isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails asp net core and visual studio information disclosure vulnerability vulnerabilityurl
| 0
|
19,133
| 25,187,442,620
|
IssuesEvent
|
2022-11-11 19:35:19
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
opened
|
Páginas extras coletadas ao coletar arquivos através de navegação do browser
|
[1] Bug [0] Desenvolvimento [2] Média Prioridade [3] Processamento Dinâmico
|
## Comportamento Esperado
Quando clicamos em um link (ou qualquer ação similar) que leve ao download de um arquivo no mecanismo de passos, desejamos que o arquivo seja baixado sem gerar nenhuma página correspondente.
## Comportamento Atual
Ao gerar um download de arquivo através de navegação no mecanismo de passos, o arquivo é baixado corretamente, mas uma página vazia correspondente também é criada.
## Passos para reproduzir o erro
Executar a coleta disponibilizada abaixo, e verificar que na página temos 8 links, dos quais 7 levam a outras páginas e 1 leva a um arquivo. Após a execução, verificar que o arquivo é baixado corretamente, mas que 9 páginas foram salvas. O esperado seriam 8 (a página inicial + os 7 links). É possivel identificar que uma das páginas está vazia, e portanto deve corresponder ao arquivo baixado.
## Especificações da Coleta
```
{
"source_name": "[Realiza\u00e7\u00e3o-F01] Coleta de Informa\u00e7\u00f5es_Intitucionais_do_Munic\u00edpio - Leis_Municipais de Santana dos Montes",
"base_url": "http:\/\/pt.santanadosmontes.mg.gov.br\/Legislacao",
"obey_robots": false,
"ignore_data_crawled_in_previous_instances": false,
"crawler_description": "a",
"crawler_type_desc": "Contratos",
"crawler_issue": 0,
"data_path": "santana_dos_montes_leis_municipais",
"sc_scheduler_persist": true,
"sc_scheduler_queue_refresh": 10,
"sc_queue_hits": 10,
"sc_queue_window": 60,
"sc_queue_moderated": true,
"sc_dupefilter_timeout": 600,
"sc_global_page_per_domain_limit": null,
"sc_global_page_per_domain_limit_timeout": 600,
"sc_domain_max_page_timeout": 600,
"sc_scheduler_ip_refresh": 60,
"sc_scheduler_backlog_blacklist": true,
"sc_scheduler_type_enabled": true,
"sc_scheduler_ip_enabled": true,
"sc_scheduler_item_retries": 3,
"sc_scheduler_queue_timeout": 3600,
"sc_httperror_allow_all": true,
"sc_retry_times": 3,
"sc_download_timeout": 10,
"antiblock_download_delay": 2,
"antiblock_autothrottle_enabled": false,
"antiblock_autothrottle_start_delay": 2,
"antiblock_autothrottle_max_delay": 10,
"antiblock_ip_rotation_enabled": false,
"antiblock_ip_rotation_type": "tor",
"antiblock_max_reqs_per_ip": 10,
"antiblock_max_reuse_rounds": 10,
"antiblock_proxy_list": "",
"antiblock_user_agent_rotation_enabled": false,
"antiblock_reqs_per_user_agent": 100,
"antiblock_user_agents_list": "",
"antiblock_insert_cookies_enabled": false,
"antiblock_cookies_list": "",
"captcha": "none",
"has_webdriver": false,
"webdriver_path": null,
"img_xpath": null,
"sound_xpath": null,
"dynamic_processing": true,
"browser_type": "chromium",
"skip_iter_errors": true,
"browser_resolution_width": 1280,
"browser_resolution_height": 720,
"explore_links": false,
"link_extractor_max_depth": 1,
"link_extractor_allow_url": null,
"link_extractor_allow_domains": null,
"link_extractor_tags": null,
"link_extractor_attrs": null,
"link_extractor_check_type": false,
"link_extractor_process_value": "",
"download_files": false,
"download_files_allow_url": null,
"download_files_allow_extensions": ".pdf",
"download_files_allow_domains": null,
"download_files_tags": null,
"download_files_attrs": null,
"download_files_process_value": "",
"download_files_check_large_content": true,
"download_imgs": false,
"steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"salva_pagina\",\"depth\":1,\"arguments\":{}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"link\",\"children\":[{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"link\"}},{\"step\":\"abrir_em_nova_aba\",\"depth\":2,\"link_xpath\":\"link\",\"children\":[]},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"salva_pagina\",\"depth\":2,\"arguments\":{}},{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"\\\"p\u00e1gina salva\\\"\"}},{\"step\":\"fechar_aba\",\"depth\":2,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"\\\"\/\/tr[*]\/td\/span\/a\\\"\"}}}}]}",
"encoding_detection_method": 1,
"expected_runtime_category": "medium",
"templated_url_parameter_handlers": [],
"templated_url_response_handlers": [],
"instance_id": "166819348934970",
"crawler_id": 52
}
```
|
1.0
|
Páginas extras coletadas ao coletar arquivos através de navegação do browser - ## Comportamento Esperado
Quando clicamos em um link (ou qualquer ação similar) que leve ao download de um arquivo no mecanismo de passos, desejamos que o arquivo seja baixado sem gerar nenhuma página correspondente.
## Comportamento Atual
Ao gerar um download de arquivo através de navegação no mecanismo de passos, o arquivo é baixado corretamente, mas uma página vazia correspondente também é criada.
## Passos para reproduzir o erro
Executar a coleta disponibilizada abaixo, e verificar que na página temos 8 links, dos quais 7 levam a outras páginas e 1 leva a um arquivo. Após a execução, verificar que o arquivo é baixado corretamente, mas que 9 páginas foram salvas. O esperado seriam 8 (a página inicial + os 7 links). É possivel identificar que uma das páginas está vazia, e portanto deve corresponder ao arquivo baixado.
## Especificações da Coleta
```
{
"source_name": "[Realiza\u00e7\u00e3o-F01] Coleta de Informa\u00e7\u00f5es_Intitucionais_do_Munic\u00edpio - Leis_Municipais de Santana dos Montes",
"base_url": "http:\/\/pt.santanadosmontes.mg.gov.br\/Legislacao",
"obey_robots": false,
"ignore_data_crawled_in_previous_instances": false,
"crawler_description": "a",
"crawler_type_desc": "Contratos",
"crawler_issue": 0,
"data_path": "santana_dos_montes_leis_municipais",
"sc_scheduler_persist": true,
"sc_scheduler_queue_refresh": 10,
"sc_queue_hits": 10,
"sc_queue_window": 60,
"sc_queue_moderated": true,
"sc_dupefilter_timeout": 600,
"sc_global_page_per_domain_limit": null,
"sc_global_page_per_domain_limit_timeout": 600,
"sc_domain_max_page_timeout": 600,
"sc_scheduler_ip_refresh": 60,
"sc_scheduler_backlog_blacklist": true,
"sc_scheduler_type_enabled": true,
"sc_scheduler_ip_enabled": true,
"sc_scheduler_item_retries": 3,
"sc_scheduler_queue_timeout": 3600,
"sc_httperror_allow_all": true,
"sc_retry_times": 3,
"sc_download_timeout": 10,
"antiblock_download_delay": 2,
"antiblock_autothrottle_enabled": false,
"antiblock_autothrottle_start_delay": 2,
"antiblock_autothrottle_max_delay": 10,
"antiblock_ip_rotation_enabled": false,
"antiblock_ip_rotation_type": "tor",
"antiblock_max_reqs_per_ip": 10,
"antiblock_max_reuse_rounds": 10,
"antiblock_proxy_list": "",
"antiblock_user_agent_rotation_enabled": false,
"antiblock_reqs_per_user_agent": 100,
"antiblock_user_agents_list": "",
"antiblock_insert_cookies_enabled": false,
"antiblock_cookies_list": "",
"captcha": "none",
"has_webdriver": false,
"webdriver_path": null,
"img_xpath": null,
"sound_xpath": null,
"dynamic_processing": true,
"browser_type": "chromium",
"skip_iter_errors": true,
"browser_resolution_width": 1280,
"browser_resolution_height": 720,
"explore_links": false,
"link_extractor_max_depth": 1,
"link_extractor_allow_url": null,
"link_extractor_allow_domains": null,
"link_extractor_tags": null,
"link_extractor_attrs": null,
"link_extractor_check_type": false,
"link_extractor_process_value": "",
"download_files": false,
"download_files_allow_url": null,
"download_files_allow_extensions": ".pdf",
"download_files_allow_domains": null,
"download_files_tags": null,
"download_files_attrs": null,
"download_files_process_value": "",
"download_files_check_large_content": true,
"download_imgs": false,
"steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"salva_pagina\",\"depth\":1,\"arguments\":{}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"link\",\"children\":[{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"link\"}},{\"step\":\"abrir_em_nova_aba\",\"depth\":2,\"link_xpath\":\"link\",\"children\":[]},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"salva_pagina\",\"depth\":2,\"arguments\":{}},{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"\\\"p\u00e1gina salva\\\"\"}},{\"step\":\"fechar_aba\",\"depth\":2,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"\\\"\/\/tr[*]\/td\/span\/a\\\"\"}}}}]}",
"encoding_detection_method": 1,
"expected_runtime_category": "medium",
"templated_url_parameter_handlers": [],
"templated_url_response_handlers": [],
"instance_id": "166819348934970",
"crawler_id": 52
}
```
|
process
|
páginas extras coletadas ao coletar arquivos através de navegação do browser comportamento esperado quando clicamos em um link ou qualquer ação similar que leve ao download de um arquivo no mecanismo de passos desejamos que o arquivo seja baixado sem gerar nenhuma página correspondente comportamento atual ao gerar um download de arquivo através de navegação no mecanismo de passos o arquivo é baixado corretamente mas uma página vazia correspondente também é criada passos para reproduzir o erro executar a coleta disponibilizada abaixo e verificar que na página temos links dos quais levam a outras páginas e leva a um arquivo após a execução verificar que o arquivo é baixado corretamente mas que páginas foram salvas o esperado seriam a página inicial os links é possivel identificar que uma das páginas está vazia e portanto deve corresponder ao arquivo baixado especificações da coleta source name coleta de informa intitucionais do munic leis municipais de santana dos montes base url http pt santanadosmontes mg gov br legislacao obey robots false ignore data crawled in previous instances false crawler description a crawler type desc contratos crawler issue data path santana dos montes leis municipais sc scheduler persist true sc scheduler queue refresh sc queue hits sc queue window sc queue moderated true sc dupefilter timeout sc global page per domain limit null sc global page per domain limit timeout sc domain max page timeout sc scheduler ip refresh sc scheduler backlog blacklist true sc scheduler type enabled true sc scheduler ip enabled true sc scheduler item retries sc scheduler queue timeout sc httperror allow all true sc retry times sc download timeout antiblock download delay antiblock autothrottle enabled false antiblock autothrottle start delay antiblock autothrottle max delay antiblock ip rotation enabled false antiblock ip rotation type tor antiblock max reqs per ip antiblock max reuse rounds antiblock proxy list antiblock user agent rotation enabled false antiblock reqs per user agent antiblock user agents list antiblock insert cookies enabled false antiblock cookies list captcha none has webdriver false webdriver path null img xpath null sound xpath null dynamic processing true browser type chromium skip iter errors true browser resolution width browser resolution height explore links false link extractor max depth link extractor allow url null link extractor allow domains null link extractor tags null link extractor attrs null link extractor check type false link extractor process value download files false download files allow url null download files allow extensions pdf download files allow domains null download files tags null download files attrs null download files process value download files check large content true download imgs false steps step root depth children step espere depth arguments segundos step salva pagina depth arguments step imprime depth arguments texto p salva step fechar aba depth arguments iterable call step localiza elementos arguments xpath tr td span a encoding detection method expected runtime category medium templated url parameter handlers templated url response handlers instance id crawler id
| 1
|
1,978
| 4,805,290,972
|
IssuesEvent
|
2016-11-02 15:42:11
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
A bug displaying boolean value
|
bug inprocess
|
Hi,
I had some data like [{test:true},{test:true},{test:false}]
The True value could display well, but the false value was empty.
Could you check this?
Best,
Yang
|
1.0
|
A bug displaying boolean value - Hi,
I had some data like [{test:true},{test:true},{test:false}]
The True value could display well, but the false value was empty.
Could you check this?
Best,
Yang
|
process
|
a bug displaying boolean value hi i had some data like the true value could display well but the false value was empty could you check this best yang
| 1
|
100,546
| 30,723,069,200
|
IssuesEvent
|
2023-07-27 17:22:47
|
spack/spack
|
https://api.github.com/repos/spack/spack
|
opened
|
Installation issue: cgns tries to use mpifc even if fortran is disabled
|
build-error
|
### Steps to reproduce the issue
```console
$ spack spec -I cgns@4.3.0~fortran ^mpich@4.1.1~fortran
Input spec
--------------------------------
- cgns@4.3.0~fortran
- ^mpich@4.1.1~fortran
Concretized
--------------------------------
- cgns@4.3.0%clang@16.0.0~base_scope~fortran+hdf5~int64~ipo~legacy~mem_debug+mpi+scoping+shared~static~testing~tools build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell
[+] ^cmake@3.25.1%clang@16.0.0~doc+ncurses+ownlibs~qt build_system=generic build_type=Release arch=linux-ubuntu20.04-haswell
[+] ^gmake@4.2.1%clang@16.0.0~guile build_system=autotools patches=ca60bd9,fe5b60d arch=linux-ubuntu20.04-haswell
[+] ^hdf5@1.12.2%clang@16.0.0~cxx~fortran~hl~ipo~java+mpi+shared~szip~threadsafe+tools api=default build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell
[+] ^pkgconf@1.8.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^zlib@1.2.13%clang@16.0.0+optimize+pic+shared build_system=makefile arch=linux-ubuntu20.04-haswell
[+] ^mpich@4.1.1%clang@16.0.0~argobots~cuda~fortran+hwloc+hydra+libxml2+pci~rocm+romio~slurm~two_level_namespace~vci~verbs+wrapperrpath build_system=autotools datatype-engine=auto device=ch4 netmod=ofi pmi=pmi arch=linux-ubuntu20.04-haswell
[+] ^findutils@4.7.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^hwloc@2.9.1%clang@16.0.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell
[+] ^ncurses@6.2.20200212%clang@16.0.0+symlinks+termlib abi=6 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libfabric@1.18.0%clang@16.0.0~debug~kdreg build_system=autotools fabrics=sockets,tcp,udp arch=linux-ubuntu20.04-haswell
[+] ^libpciaccess@0.17%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libtool@2.4.6%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^util-macros@1.19.3%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libxml2@2.10.3%clang@16.0.0~python build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libiconv@1.17%clang@16.0.0 build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell
[+] ^xz@5.2.4%clang@16.0.0~pic build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell
[+] ^yaksa@0.2%gcc@12.2.0~cuda~rocm build_system=autotools arch=linux-ubuntu20.04-haswell
[-] ^autoconf@2.69%gcc@12.2.0 build_system=autotools patches=7793209 arch=linux-ubuntu20.04-haswell
[-] ^automake@1.16.1%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[-] ^libtool@2.4.6%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[-] ^m4@1.4.18%gcc@12.2.0+sigsegv build_system=autotools patches=3877ab5,fc9b616 arch=linux-ubuntu20.04-haswell
[-] ^python@3.8.10%gcc@12.2.0+bz2+crypt+ctypes+dbm~debug+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tkinter+uuid+zlib build_system=generic patches=0d98e93,4c24573,f2fd060 arch=linux-ubuntu20.04-haswell
```
### Error message
<details><summary>Error message</summary>
<pre>
==> Error: AttributeError: 'Spec' object has no attribute 'mpifc'
The 'cgns' package cannot find an attribute while trying to build from sources. This might be due to a change in Spack's package format to support multiple build-systems for a single package. You can fix this by updating the build recipe, and you can also report the issue as a bug. More information at https://spack.readthedocs.io/en/latest/packaging_guide.html#installation-procedure
/home/nmmoral/opt/spack/var/spack/repos/builtin/packages/cgns/package.py:93, in cmake_args:
90 [
91 "-DCMAKE_C_COMPILER=%s" % spec["mpi"].mpicc,
92 "-DCMAKE_CXX_COMPILER=%s" % spec["mpi"].mpicxx,
>> 93 "-DCMAKE_Fortran_COMPILER=%s" % spec["mpi"].mpifc,
94 ]
95 )
96
See build log for details:
/tmp/nmmoral/spack-stage/spack-stage-cgns-4.3.0-h42e75qpcnab3ozroegcqbrx2ghlqkat/spack-build-out.txt
</pre></details>
### Information on your system
```
* **Spack:** 0.20.0.dev0 (f5c32d57e051a8ea6f04bc86f3df5867ea1ab49e)
* **Python:** 3.8.10
* **Platform:** linux-ubuntu20.04-skylake_avx512
* **Concretizer:** clingo
```
### Additional information
Maintainer: @gsjaardema
Note the bug here is that the recipe for cgns unconditionally sets the fortran compiler to mpifc even if mpifc is not available in the mpi install.
spack-build-out.txt
```
==> cgns: Executing phase: 'cmake'
```
I can't upload my spack-build-env but please contact me if you need some specific info. I don't think it's too relevant to the bug however.
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
1.0
|
Installation issue: cgns tries to use mpifc even if fortran is disabled - ### Steps to reproduce the issue
```console
$ spack spec -I cgns@4.3.0~fortran ^mpich@4.1.1~fortran
Input spec
--------------------------------
- cgns@4.3.0~fortran
- ^mpich@4.1.1~fortran
Concretized
--------------------------------
- cgns@4.3.0%clang@16.0.0~base_scope~fortran+hdf5~int64~ipo~legacy~mem_debug+mpi+scoping+shared~static~testing~tools build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell
[+] ^cmake@3.25.1%clang@16.0.0~doc+ncurses+ownlibs~qt build_system=generic build_type=Release arch=linux-ubuntu20.04-haswell
[+] ^gmake@4.2.1%clang@16.0.0~guile build_system=autotools patches=ca60bd9,fe5b60d arch=linux-ubuntu20.04-haswell
[+] ^hdf5@1.12.2%clang@16.0.0~cxx~fortran~hl~ipo~java+mpi+shared~szip~threadsafe+tools api=default build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell
[+] ^pkgconf@1.8.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^zlib@1.2.13%clang@16.0.0+optimize+pic+shared build_system=makefile arch=linux-ubuntu20.04-haswell
[+] ^mpich@4.1.1%clang@16.0.0~argobots~cuda~fortran+hwloc+hydra+libxml2+pci~rocm+romio~slurm~two_level_namespace~vci~verbs+wrapperrpath build_system=autotools datatype-engine=auto device=ch4 netmod=ofi pmi=pmi arch=linux-ubuntu20.04-haswell
[+] ^findutils@4.7.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^hwloc@2.9.1%clang@16.0.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell
[+] ^ncurses@6.2.20200212%clang@16.0.0+symlinks+termlib abi=6 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libfabric@1.18.0%clang@16.0.0~debug~kdreg build_system=autotools fabrics=sockets,tcp,udp arch=linux-ubuntu20.04-haswell
[+] ^libpciaccess@0.17%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libtool@2.4.6%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^util-macros@1.19.3%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libxml2@2.10.3%clang@16.0.0~python build_system=autotools arch=linux-ubuntu20.04-haswell
[+] ^libiconv@1.17%clang@16.0.0 build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell
[+] ^xz@5.2.4%clang@16.0.0~pic build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell
[+] ^yaksa@0.2%gcc@12.2.0~cuda~rocm build_system=autotools arch=linux-ubuntu20.04-haswell
[-] ^autoconf@2.69%gcc@12.2.0 build_system=autotools patches=7793209 arch=linux-ubuntu20.04-haswell
[-] ^automake@1.16.1%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[-] ^libtool@2.4.6%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell
[-] ^m4@1.4.18%gcc@12.2.0+sigsegv build_system=autotools patches=3877ab5,fc9b616 arch=linux-ubuntu20.04-haswell
[-] ^python@3.8.10%gcc@12.2.0+bz2+crypt+ctypes+dbm~debug+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tkinter+uuid+zlib build_system=generic patches=0d98e93,4c24573,f2fd060 arch=linux-ubuntu20.04-haswell
```
### Error message
<details><summary>Error message</summary>
<pre>
==> Error: AttributeError: 'Spec' object has no attribute 'mpifc'
The 'cgns' package cannot find an attribute while trying to build from sources. This might be due to a change in Spack's package format to support multiple build-systems for a single package. You can fix this by updating the build recipe, and you can also report the issue as a bug. More information at https://spack.readthedocs.io/en/latest/packaging_guide.html#installation-procedure
/home/nmmoral/opt/spack/var/spack/repos/builtin/packages/cgns/package.py:93, in cmake_args:
90 [
91 "-DCMAKE_C_COMPILER=%s" % spec["mpi"].mpicc,
92 "-DCMAKE_CXX_COMPILER=%s" % spec["mpi"].mpicxx,
>> 93 "-DCMAKE_Fortran_COMPILER=%s" % spec["mpi"].mpifc,
94 ]
95 )
96
See build log for details:
/tmp/nmmoral/spack-stage/spack-stage-cgns-4.3.0-h42e75qpcnab3ozroegcqbrx2ghlqkat/spack-build-out.txt
</pre></details>
### Information on your system
```
* **Spack:** 0.20.0.dev0 (f5c32d57e051a8ea6f04bc86f3df5867ea1ab49e)
* **Python:** 3.8.10
* **Platform:** linux-ubuntu20.04-skylake_avx512
* **Concretizer:** clingo
```
### Additional information
Maintainer: @gsjaardema
Note the bug here is that the recipe for cgns unconditionally sets the fortran compiler to mpifc even if mpifc is not available in the mpi install.
spack-build-out.txt
```
==> cgns: Executing phase: 'cmake'
```
I can't upload my spack-build-env but please contact me if you need some specific info. I don't think it's too relevant to the bug however.
### General information
- [X] I have run `spack debug report` and reported the version of Spack/Python/Platform
- [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers
- [X] I have uploaded the build log and environment files
- [X] I have searched the issues of this repo and believe this is not a duplicate
|
non_process
|
installation issue cgns tries to use mpifc even if fortran is disabled steps to reproduce the issue console spack spec i cgns fortran mpich fortran input spec cgns fortran mpich fortran concretized cgns clang base scope fortran ipo legacy mem debug mpi scoping shared static testing tools build system cmake build type relwithdebinfo generator make arch linux haswell cmake clang doc ncurses ownlibs qt build system generic build type release arch linux haswell gmake clang guile build system autotools patches arch linux haswell clang cxx fortran hl ipo java mpi shared szip threadsafe tools api default build system cmake build type relwithdebinfo generator make arch linux haswell pkgconf clang build system autotools arch linux haswell zlib clang optimize pic shared build system makefile arch linux haswell mpich clang argobots cuda fortran hwloc hydra pci rocm romio slurm two level namespace vci verbs wrapperrpath build system autotools datatype engine auto device netmod ofi pmi pmi arch linux haswell findutils clang build system autotools arch linux haswell hwloc clang cairo cuda gl libudev netloc nvml oneapi level zero opencl pci rocm build system autotools libs shared static arch linux haswell ncurses clang symlinks termlib abi build system autotools arch linux haswell libfabric clang debug kdreg build system autotools fabrics sockets tcp udp arch linux haswell libpciaccess clang build system autotools arch linux haswell libtool clang build system autotools arch linux haswell util macros clang build system autotools arch linux haswell clang python build system autotools arch linux haswell libiconv clang build system autotools libs shared static arch linux haswell xz clang pic build system autotools libs shared static arch linux haswell yaksa gcc cuda rocm build system autotools arch linux haswell autoconf gcc build system autotools patches arch linux haswell automake gcc build system autotools arch linux haswell libtool gcc build system autotools arch linux haswell gcc sigsegv build system autotools patches arch linux haswell python gcc crypt ctypes dbm debug lzma nis optimizations pic pyexpat pythoncmd readline shared ssl tkinter uuid zlib build system generic patches arch linux haswell error message error message error attributeerror spec object has no attribute mpifc the cgns package cannot find an attribute while trying to build from sources this might be due to a change in spack s package format to support multiple build systems for a single package you can fix this by updating the build recipe and you can also report the issue as a bug more information at home nmmoral opt spack var spack repos builtin packages cgns package py in cmake args dcmake c compiler s spec mpicc dcmake cxx compiler s spec mpicxx dcmake fortran compiler s spec mpifc see build log for details tmp nmmoral spack stage spack stage cgns spack build out txt information on your system spack python platform linux skylake concretizer clingo additional information maintainer gsjaardema note the bug here is that the recipe for cgns unconditionally sets the fortran compiler to mpifc even if mpifc is not available in the mpi install spack build out txt cgns executing phase cmake i can t upload my spack build env but please contact me if you need some specific info i don t think it s too relevant to the bug however general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
| 0
|
1,940
| 4,769,390,751
|
IssuesEvent
|
2016-10-26 12:26:27
|
openvstorage/alba
|
https://api.github.com/repos/openvstorage/alba
|
closed
|
Alba proxy crashes with Unix_error "Invalid argument" select
|
priority_critical process_wontfix type_bug
|
```
alba: internal error, uncaught exception:
(Unix.Unix_error "Invalid argument" select "")
Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50
Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39
Called from file "src/unix/lwt_main.ml", line 41, characters 8-82
Uncaught exception:
(Unix.Unix_error "Invalid argument" select "")
Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50
Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39
Called from file "src/unix/lwt_main.ml", line 41, characters 8-82
Called from file "pervasives.ml", line 482, characters 30-33
Called from file "pervasives.ml", line 487, characters 2-15
Called from file "src/alba.ml", line 828, characters 16-22
```
Saw this 450 times in a log file of a day.
The env (OVH) was not configured to use rdma.
|
1.0
|
Alba proxy crashes with Unix_error "Invalid argument" select - ```
alba: internal error, uncaught exception:
(Unix.Unix_error "Invalid argument" select "")
Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50
Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39
Called from file "src/unix/lwt_main.ml", line 41, characters 8-82
Uncaught exception:
(Unix.Unix_error "Invalid argument" select "")
Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50
Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39
Called from file "src/unix/lwt_main.ml", line 41, characters 8-82
Called from file "pervasives.ml", line 482, characters 30-33
Called from file "pervasives.ml", line 487, characters 2-15
Called from file "src/alba.ml", line 828, characters 16-22
```
Saw this 450 times in a log file of a day.
The env (OVH) was not configured to use rdma.
|
process
|
alba proxy crashes with unix error invalid argument select alba internal error uncaught exception unix unix error invalid argument select raised by primitive operation at file lwt rsocket ml line characters called from file src unix lwt engine ml line characters called from file src unix lwt main ml line characters uncaught exception unix unix error invalid argument select raised by primitive operation at file lwt rsocket ml line characters called from file src unix lwt engine ml line characters called from file src unix lwt main ml line characters called from file pervasives ml line characters called from file pervasives ml line characters called from file src alba ml line characters saw this times in a log file of a day the env ovh was not configured to use rdma
| 1
|
18,899
| 6,655,669,744
|
IssuesEvent
|
2017-09-29 17:22:42
|
habitat-sh/habitat
|
https://api.github.com/repos/habitat-sh/habitat
|
opened
|
Members tab should only be visible to origin members
|
A-builder C-bug L-javascript
|
E.g.: https://bldr.habitat.sh/#/origins/cnunciato/members
The invitation API calls do fail, of course, but the UI should make it clear you really shouldn't be here.
|
1.0
|
Members tab should only be visible to origin members - E.g.: https://bldr.habitat.sh/#/origins/cnunciato/members
The invitation API calls do fail, of course, but the UI should make it clear you really shouldn't be here.
|
non_process
|
members tab should only be visible to origin members e g the invitation api calls do fail of course but the ui should make it clear you really shouldn t be here
| 0
|
16,199
| 20,710,470,723
|
IssuesEvent
|
2022-03-12 00:04:40
|
scikit-learn/scikit-learn
|
https://api.github.com/repos/scikit-learn/scikit-learn
|
closed
|
_handle_zeros_in_scale causing improper scaling when using StandardScaler()
|
Bug module:preprocessing
|
<!--
Before submitting a bug, please make sure the issue hasn't been already
addressed by searching through the past issues.
-->
#### Describe the bug
There is no floating point tolerance in function _handle_zeros_in_scale for checking if scale == 0.0. As a result, floating point precision can cause this check to incorrectly fail and not set scale to 1.0. The end result is to potentially have an incorrectly scaled values when using StandardScaler() since the value of scale_ will be near 0 instead of 1, introducing numerical instability.
#### Steps/Code to Reproduce
```
from sklearn.preprocessing import StandardScaler
import numpy as np
data_fails = np.full((1000, 1), 14.62, dtype=float).reshape(-1,1) #array filled with 14.62, causes issue
data_works = np.full((1000,1), 100.0 , dtype=float).reshape(-1,1) #array filled with 100.0, works as intended
scaler_fails = StandardScaler()
scaler_works = StandardScaler()
scaled_fails = scaler_fails.fit_transform(data_fails) #Returns array filled with -1.0
scaled_works = scaler_works.fit_transform(data_works) #Returns array fill with 0.0
print('\n Results: \n\n')
print(scaled_fails[0][0])
print(scaled_works[0][0])
```
#### Expected Results
Expected both scaled results to be zero vector since both are constant-valued vectors.
#### Actual Results
Standard scaling subtracts mean and divides by standard deviation when appropriate flags are set as in example above. Variance of constant valued vector is 0 which should be caught and replaced by 1 in function _handle_zeros_in_scale. However, this is not happening due variations introduced by floating point representation. Results in mean_ being divided by small floating point value resulting in incorrect scaling when using StandardScaler().
Error occurs at line number 77 in my version of _data inside function _handle_zeros_in_scale. Currently reads:
scale[scale == 0.0] = 1.0
#### Versions
Python dependencies:
pip: 20.0.2
setuptools: 47.1.1.post20200604
sklearn: 0.22.1
numpy: 1.18.1
scipy: 1.4.1
Cython: None
pandas: 1.0.3
matplotlib: 3.2.1
joblib: 0.15.1
<!-- Thanks for contributing! -->
|
1.0
|
_handle_zeros_in_scale causing improper scaling when using StandardScaler() - <!--
Before submitting a bug, please make sure the issue hasn't been already
addressed by searching through the past issues.
-->
#### Describe the bug
There is no floating point tolerance in function _handle_zeros_in_scale for checking if scale == 0.0. As a result, floating point precision can cause this check to incorrectly fail and not set scale to 1.0. The end result is to potentially have an incorrectly scaled values when using StandardScaler() since the value of scale_ will be near 0 instead of 1, introducing numerical instability.
#### Steps/Code to Reproduce
```
from sklearn.preprocessing import StandardScaler
import numpy as np
data_fails = np.full((1000, 1), 14.62, dtype=float).reshape(-1,1) #array filled with 14.62, causes issue
data_works = np.full((1000,1), 100.0 , dtype=float).reshape(-1,1) #array filled with 100.0, works as intended
scaler_fails = StandardScaler()
scaler_works = StandardScaler()
scaled_fails = scaler_fails.fit_transform(data_fails) #Returns array filled with -1.0
scaled_works = scaler_works.fit_transform(data_works) #Returns array fill with 0.0
print('\n Results: \n\n')
print(scaled_fails[0][0])
print(scaled_works[0][0])
```
#### Expected Results
Expected both scaled results to be zero vector since both are constant-valued vectors.
#### Actual Results
Standard scaling subtracts mean and divides by standard deviation when appropriate flags are set as in example above. Variance of constant valued vector is 0 which should be caught and replaced by 1 in function _handle_zeros_in_scale. However, this is not happening due variations introduced by floating point representation. Results in mean_ being divided by small floating point value resulting in incorrect scaling when using StandardScaler().
Error occurs at line number 77 in my version of _data inside function _handle_zeros_in_scale. Currently reads:
scale[scale == 0.0] = 1.0
#### Versions
Python dependencies:
pip: 20.0.2
setuptools: 47.1.1.post20200604
sklearn: 0.22.1
numpy: 1.18.1
scipy: 1.4.1
Cython: None
pandas: 1.0.3
matplotlib: 3.2.1
joblib: 0.15.1
<!-- Thanks for contributing! -->
|
process
|
handle zeros in scale causing improper scaling when using standardscaler before submitting a bug please make sure the issue hasn t been already addressed by searching through the past issues describe the bug there is no floating point tolerance in function handle zeros in scale for checking if scale as a result floating point precision can cause this check to incorrectly fail and not set scale to the end result is to potentially have an incorrectly scaled values when using standardscaler since the value of scale will be near instead of introducing numerical instability steps code to reproduce from sklearn preprocessing import standardscaler import numpy as np data fails np full dtype float reshape array filled with causes issue data works np full dtype float reshape array filled with works as intended scaler fails standardscaler scaler works standardscaler scaled fails scaler fails fit transform data fails returns array filled with scaled works scaler works fit transform data works returns array fill with print n results n n print scaled fails print scaled works expected results expected both scaled results to be zero vector since both are constant valued vectors actual results standard scaling subtracts mean and divides by standard deviation when appropriate flags are set as in example above variance of constant valued vector is which should be caught and replaced by in function handle zeros in scale however this is not happening due variations introduced by floating point representation results in mean being divided by small floating point value resulting in incorrect scaling when using standardscaler error occurs at line number in my version of data inside function handle zeros in scale currently reads scale versions python dependencies pip setuptools sklearn numpy scipy cython none pandas matplotlib joblib
| 1
|
12,423
| 14,923,397,384
|
IssuesEvent
|
2021-01-23 18:46:06
|
AcademySoftwareFoundation/OpenCue
|
https://api.github.com/repos/AcademySoftwareFoundation/OpenCue
|
opened
|
docker files need to pin max pip version for python2
|
process
|
**Describe the process**
The latest version of `pip` (21.0) and `setuptools (45) has just dropped python 2 support.
As a result building the sandbox environment fails:
```
docker-compose --project-directory . -f sandbox/docker-compose.yml build
```
```
Step 8/25 : RUN python -m pip install --upgrade setuptools
---> Running in 8eb1abc898a2
Traceback (most recent call last):
File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
File "/usr/lib/python2.7/site-packages/pip/__main__.py", line 21, in <module>
from pip._internal.cli.main import main as _main
File "/usr/lib/python2.7/site-packages/pip/_internal/cli/main.py", line 60
sys.stderr.write(f"ERROR: {exc}")
^
SyntaxError: invalid syntax
ERROR: Service 'metrics' failed to build : The command '/bin/sh -c python -m pip install --upgrade setuptools' returned a non-zero code: 1
```
We need should pin `pip` < 21 and `setuptools` < 45 during the calls to upgrade these packages in the various Dockerfiles
```
RUN python -m pip install --upgrade 'pip<21'
RUN python -m pip install --upgrade 'setuptools<45'
```
|
1.0
|
docker files need to pin max pip version for python2 - **Describe the process**
The latest version of `pip` (21.0) and `setuptools (45) has just dropped python 2 support.
As a result building the sandbox environment fails:
```
docker-compose --project-directory . -f sandbox/docker-compose.yml build
```
```
Step 8/25 : RUN python -m pip install --upgrade setuptools
---> Running in 8eb1abc898a2
Traceback (most recent call last):
File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
File "/usr/lib/python2.7/site-packages/pip/__main__.py", line 21, in <module>
from pip._internal.cli.main import main as _main
File "/usr/lib/python2.7/site-packages/pip/_internal/cli/main.py", line 60
sys.stderr.write(f"ERROR: {exc}")
^
SyntaxError: invalid syntax
ERROR: Service 'metrics' failed to build : The command '/bin/sh -c python -m pip install --upgrade setuptools' returned a non-zero code: 1
```
We need should pin `pip` < 21 and `setuptools` < 45 during the calls to upgrade these packages in the various Dockerfiles
```
RUN python -m pip install --upgrade 'pip<21'
RUN python -m pip install --upgrade 'setuptools<45'
```
|
process
|
docker files need to pin max pip version for describe the process the latest version of pip and setuptools has just dropped python support as a result building the sandbox environment fails docker compose project directory f sandbox docker compose yml build step run python m pip install upgrade setuptools running in traceback most recent call last file usr runpy py line in run module as main main fname loader pkg name file usr runpy py line in run code exec code in run globals file usr lib site packages pip main py line in from pip internal cli main import main as main file usr lib site packages pip internal cli main py line sys stderr write f error exc syntaxerror invalid syntax error service metrics failed to build the command bin sh c python m pip install upgrade setuptools returned a non zero code we need should pin pip and setuptools during the calls to upgrade these packages in the various dockerfiles run python m pip install upgrade pip run python m pip install upgrade setuptools
| 1
|
22,713
| 32,038,297,439
|
IssuesEvent
|
2023-09-22 17:05:17
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
nx 16.8.1 has 2 guarddog issues
|
npm-install-script npm-silent-process-execution
|
```{"npm-install-script":[{"code":" \"postinstall\": \"node ./bin/post-install\"","location":"package/package.json:12","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const p = (0, child_process_1.spawn)('node', [scriptPath, `\"${this.cachePath}\"`], {\n stdio: 'ignore',\n detached: true,\n shell: false,\n });","location":"package/src/tasks-runner/cache.js:28","message":"This package is silently executing another executable"}]}```
|
1.0
|
nx 16.8.1 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"postinstall\": \"node ./bin/post-install\"","location":"package/package.json:12","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const p = (0, child_process_1.spawn)('node', [scriptPath, `\"${this.cachePath}\"`], {\n stdio: 'ignore',\n detached: true,\n shell: false,\n });","location":"package/src/tasks-runner/cache.js:28","message":"This package is silently executing another executable"}]}```
|
process
|
nx has guarddog issues npm install script npm silent process execution n stdio ignore n detached true n shell false n location package src tasks runner cache js message this package is silently executing another executable
| 1
|
17,984
| 24,006,190,811
|
IssuesEvent
|
2022-09-14 14:58:22
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[processor/k8sattributes] agent passthrough, gateway k8s.pod.ip configuration broken post v0.55.0 release
|
bug priority:p2 processor/k8sattributes
|
**Describe the bug**
In an agent/gateway k8s deployment, the `k8sattributes` processor no longer correctly adds k8s metadata to the Resource associated with traces emitted by a pod elsewhere in the cluster. I suspect this may have broken due to [PR #8465](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/8465)
**Steps to reproduce**
Create a daemonset agent/deployment gateway collector setup in a k8s cluster with the configuration files shown below. Generate traces from a pod that sends to the local daemonset agent pod. Observe the debug logs on the gateway collector.
**What did you expect to see?**
Logs for each ResourceSpans indicating each Resource with a preexisting `k8s.pod.ip` attribute had the Resource augmented with `k8s.namespace.name`, `k8s.pod.uid` and `k8s.pod.name`.
**What did you see instead?**
Logs showing the incoming traces have a `Resource` with the correct pod IP in the `k8s.pod.ip` attribute, but **none** of `k8s.namespace.name`, `k8s.pod.uid` nor `k8s.pod.name`.
**What version did you use?**
Version: v0.56.0
**What config did you use?**
On the agent side:
```yaml
receivers:
otlp:
protocols:
grpc: 0.0.0.0:4317
processors:
k8sattributes:
passthrough: true
batch:
exporters:
otlp:
endpoint: otel-gateway:4317
service:
pipelines:
traces:
receivers: [ otlp ]
processors: [ k8sattributes, batch ]
exporters: [ otlp ]
```
and on the gateway side
```yaml
receivers:
otlp:
protocols:
grpc: 0.0.0.0:4317
processors:
batch:
k8sattributes:
pod_association:
- sources:
- from: resource_attribute
name: k8s.pod.ip
extract:
metadata:
- k8s.namespace.name
- k8s.pod.name
- k8s.pod.uid
exporters:
logging:
loglevel: debug
service:
pipelines:
traces:
receivers: [ otlp ]
processors: [ k8sattributes, batch ]
exporters: [ logging ]
```
**Environment**
OS: RHEL8 Linux
Compiler(if manually compiled): go1.17
**Additional context**
From the extension debug logs, the incoming Resource is already augmented by the agent pod with `k8s.pod.ip=<podIP>`.
In `kubernetesprocessor.processResource()`, the debug message `"evaluating pod identifier [ {Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podIp>" }, ..]"` is printed, but it appears to return false from `WatchClient.GetPod(<podIdentifier>)`. This seems to indicate that there is no entry in `c.Pods[]` for that specific PodIdentifier.
The primary place in the code where pods are added to the `c.Pods[]` map is in `WatchClient.addOrUpdatePod()`.
Prior to the aforementioned PR, the keys for the Pods map were Strings containing either the IP address or the pod UID, and there was no distinction between whether it came from the "connection" or from a "resource_attribute".
After the PR, the key(s) returned from`WatchClient.getIdentifiersFromAssoc()` are inserted into the Pods map, with each key pointing to the pod info. To correctly match the key which is printed in `evaluating pod identifier...`, `getIdentifiersFromAssoc()` should return the key `{Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podip>" }` when given the input association specification, namely:
```yaml
pod_association:
- sources:
- from: resource_attribute
name: k8s.pod.ip
```
Based on code inspection, it appears this does not happen. I suspect changing the [switch statement in `getIdentifiersFromAssoc()`](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/fffc27fe3a5f360289d21604e61779899ef0fefa/processor/k8sattributesprocessor/internal/kube/client.go#L468-L483) from:
```go
case conventions.AttributeHostName:
attr = pod.Address
```
to
```go
case conventions.AttributeHostName, "k8s.pod.ip":
attr = pod.Address
```
would probably fix this problem. But I have only ascertained this via code inspection, not compiling and running the code!
Possibly related: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/13119
|
1.0
|
[processor/k8sattributes] agent passthrough, gateway k8s.pod.ip configuration broken post v0.55.0 release - **Describe the bug**
In an agent/gateway k8s deployment, the `k8sattributes` processor no longer correctly adds k8s metadata to the Resource associated with traces emitted by a pod elsewhere in the cluster. I suspect this may have broken due to [PR #8465](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/8465)
**Steps to reproduce**
Create a daemonset agent/deployment gateway collector setup in a k8s cluster with the configuration files shown below. Generate traces from a pod that sends to the local daemonset agent pod. Observe the debug logs on the gateway collector.
**What did you expect to see?**
Logs for each ResourceSpans indicating each Resource with a preexisting `k8s.pod.ip` attribute had the Resource augmented with `k8s.namespace.name`, `k8s.pod.uid` and `k8s.pod.name`.
**What did you see instead?**
Logs showing the incoming traces have a `Resource` with the correct pod IP in the `k8s.pod.ip` attribute, but **none** of `k8s.namespace.name`, `k8s.pod.uid` nor `k8s.pod.name`.
**What version did you use?**
Version: v0.56.0
**What config did you use?**
On the agent side:
```yaml
receivers:
otlp:
protocols:
grpc: 0.0.0.0:4317
processors:
k8sattributes:
passthrough: true
batch:
exporters:
otlp:
endpoint: otel-gateway:4317
service:
pipelines:
traces:
receivers: [ otlp ]
processors: [ k8sattributes, batch ]
exporters: [ otlp ]
```
and on the gateway side
```yaml
receivers:
otlp:
protocols:
grpc: 0.0.0.0:4317
processors:
batch:
k8sattributes:
pod_association:
- sources:
- from: resource_attribute
name: k8s.pod.ip
extract:
metadata:
- k8s.namespace.name
- k8s.pod.name
- k8s.pod.uid
exporters:
logging:
loglevel: debug
service:
pipelines:
traces:
receivers: [ otlp ]
processors: [ k8sattributes, batch ]
exporters: [ logging ]
```
**Environment**
OS: RHEL8 Linux
Compiler(if manually compiled): go1.17
**Additional context**
From the extension debug logs, the incoming Resource is already augmented by the agent pod with `k8s.pod.ip=<podIP>`.
In `kubernetesprocessor.processResource()`, the debug message `"evaluating pod identifier [ {Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podIp>" }, ..]"` is printed, but it appears to return false from `WatchClient.GetPod(<podIdentifier>)`. This seems to indicate that there is no entry in `c.Pods[]` for that specific PodIdentifier.
The primary place in the code where pods are added to the `c.Pods[]` map is in `WatchClient.addOrUpdatePod()`.
Prior to the aforementioned PR, the keys for the Pods map were Strings containing either the IP address or the pod UID, and there was no distinction between whether it came from the "connection" or from a "resource_attribute".
After the PR, the key(s) returned from`WatchClient.getIdentifiersFromAssoc()` are inserted into the Pods map, with each key pointing to the pod info. To correctly match the key which is printed in `evaluating pod identifier...`, `getIdentifiersFromAssoc()` should return the key `{Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podip>" }` when given the input association specification, namely:
```yaml
pod_association:
- sources:
- from: resource_attribute
name: k8s.pod.ip
```
Based on code inspection, it appears this does not happen. I suspect changing the [switch statement in `getIdentifiersFromAssoc()`](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/fffc27fe3a5f360289d21604e61779899ef0fefa/processor/k8sattributesprocessor/internal/kube/client.go#L468-L483) from:
```go
case conventions.AttributeHostName:
attr = pod.Address
```
to
```go
case conventions.AttributeHostName, "k8s.pod.ip":
attr = pod.Address
```
would probably fix this problem. But I have only ascertained this via code inspection, not compiling and running the code!
Possibly related: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/13119
|
process
|
agent passthrough gateway pod ip configuration broken post release describe the bug in an agent gateway deployment the processor no longer correctly adds metadata to the resource associated with traces emitted by a pod elsewhere in the cluster i suspect this may have broken due to steps to reproduce create a daemonset agent deployment gateway collector setup in a cluster with the configuration files shown below generate traces from a pod that sends to the local daemonset agent pod observe the debug logs on the gateway collector what did you expect to see logs for each resourcespans indicating each resource with a preexisting pod ip attribute had the resource augmented with namespace name pod uid and pod name what did you see instead logs showing the incoming traces have a resource with the correct pod ip in the pod ip attribute but none of namespace name pod uid nor pod name what version did you use version what config did you use on the agent side yaml receivers otlp protocols grpc processors passthrough true batch exporters otlp endpoint otel gateway service pipelines traces receivers processors exporters and on the gateway side yaml receivers otlp protocols grpc processors batch pod association sources from resource attribute name pod ip extract metadata namespace name pod name pod uid exporters logging loglevel debug service pipelines traces receivers processors exporters environment os linux compiler if manually compiled additional context from the extension debug logs the incoming resource is already augmented by the agent pod with pod ip in kubernetesprocessor processresource the debug message evaluating pod identifier is printed but it appears to return false from watchclient getpod this seems to indicate that there is no entry in c pods for that specific podidentifier the primary place in the code where pods are added to the c pods map is in watchclient addorupdatepod prior to the aforementioned pr the keys for the pods map were strings containing either the ip address or the pod uid and there was no distinction between whether it came from the connection or from a resource attribute after the pr the key s returned from watchclient getidentifiersfromassoc are inserted into the pods map with each key pointing to the pod info to correctly match the key which is printed in evaluating pod identifier getidentifiersfromassoc should return the key source from resource attribute name pod ip value when given the input association specification namely yaml pod association sources from resource attribute name pod ip based on code inspection it appears this does not happen i suspect changing the from go case conventions attributehostname attr pod address to go case conventions attributehostname pod ip attr pod address would probably fix this problem but i have only ascertained this via code inspection not compiling and running the code possibly related
| 1
|
15,149
| 18,906,310,222
|
IssuesEvent
|
2021-11-16 09:27:42
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
NTR: Negative Regulation of CRISPR-cas system
|
multi-species process
|
New Term: Negative regulation of CRISPR-cas system
Definition: Negative regulation of an adaptive immune response of bacteria that serves to clear host cells of foreign DNA and RNA. It has three distinct stage: acquisition of foreign DNA by integration into CRISPR loci in the host chromosome, CRISPR RNA (crRNA) biogenesis, and target interference. CISPR stands for Clustered Regularly Interspaced Short Palindromic Repeat, which describes the nature of the loci.
Notes: This reference describes the function of anti-CRISPR proteins within a bacterial genome to protect itself from its CRISPR system.
PMID: 30190308, 30190307
@sandyl27 @jimhu-tamu
This is for CACAO
|
1.0
|
NTR: Negative Regulation of CRISPR-cas system - New Term: Negative regulation of CRISPR-cas system
Definition: Negative regulation of an adaptive immune response of bacteria that serves to clear host cells of foreign DNA and RNA. It has three distinct stage: acquisition of foreign DNA by integration into CRISPR loci in the host chromosome, CRISPR RNA (crRNA) biogenesis, and target interference. CISPR stands for Clustered Regularly Interspaced Short Palindromic Repeat, which describes the nature of the loci.
Notes: This reference describes the function of anti-CRISPR proteins within a bacterial genome to protect itself from its CRISPR system.
PMID: 30190308, 30190307
@sandyl27 @jimhu-tamu
This is for CACAO
|
process
|
ntr negative regulation of crispr cas system new term negative regulation of crispr cas system definition negative regulation of an adaptive immune response of bacteria that serves to clear host cells of foreign dna and rna it has three distinct stage acquisition of foreign dna by integration into crispr loci in the host chromosome crispr rna crrna biogenesis and target interference cispr stands for clustered regularly interspaced short palindromic repeat which describes the nature of the loci notes this reference describes the function of anti crispr proteins within a bacterial genome to protect itself from its crispr system pmid jimhu tamu this is for cacao
| 1
|
10,684
| 13,464,983,681
|
IssuesEvent
|
2020-09-09 20:05:40
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
Vagrant Cloud post-processor upload errors with unexpected EOF
|
bug post-processor/vagrant-cloud
|
#### Overview of the Issue
tl;dr: Vagrant Cloud post-processor fails with an unexpected EOF when uploading a successful build targeting Virtual Box. The build is running on a Jenkins agent. I'm running the build locally at the moment, and will report back with results when it finishes.
I'm using a near-minimal Packer template with the Vagrant builder to run an Ansible build on the `bento/ubuntu-18.04` base box. Template is attached below. I've stripped it down to the bare essentials and reduced the Ansible provisioning to a series of package installs with `apt` to attempt to pinpoint the error, so there aren't many distractions.
#### Reproduction Steps
Trigger the build on a Jenkins agent via Blue Ocean pipelines.
See infrastructure details below for details.
### Packer version
Packer version `1.4.4`.
### Simplified Packer Buildfile
<details>
<summary><b>Packer template</b> (`UbuntuVM.json`)</summary>
<pre>
{
"variables": {
"cloud_token": "{{ env `VAGRANT_CLOUD_TOKEN` }}",
"version": "1.0.{{timestamp}}"
},
"builders": [{
"type": "vagrant",
"communicator": "ssh",
"provider": "virtualbox",
"source_path": "bento/ubuntu-18.04",
"add_force": true
}],
"provisioners": [{
"type": "ansible",
"playbook_file": "../provisioners/main.yml"
}],
"post-processors": [
{
"type": "vagrant-cloud",
"box_tag": "cybersecurity/UbuntuVM",
"access_token": "{{user `VAGRANT_CLOUD_TOKEN`}}",
"version": "{{user `version`}}"
}
]
}
</pre>
</details>
<details>
<summary><b>Jenkinsfile</b></summary>
<pre>
pipeline {
agent any
stages {
stage('Build') {
steps {
sh(label: 'Check directory', script: 'pwd && whoami')
sh(label: 'Verify Packer is on $PATH', script: '/usr/local/bin/packer --version')
}
}
stage('Packer') {
steps {
echo 'Packing...'
sh(label: 'Packer Build Stage!', script: '/usr/local/bin/packer build -debug UbuntuVM/packer/UbuntuVM.json')
}
}
stage('Deploy') {
steps {
echo 'Deploying...'
}
}
}
environment {
VAGRANT_CLOUD_TOKEN = 'Omitted Cause...Secrets'
}
}
</pre>
</details>
### Operating system and Environment details
**Infrastructure**
- Jenkins v2.203 running on Ubuntu 18.04, deployed to an Azure `Standard_D4S_v3` instance
- The Ubuntu instance is unmodified _except_ for the following installations:
- Vagrant v2.0.2
- Ansible v2.9.0
- Packer v1.4.4
**Build Details**
- The build is described in a declarative `Jenkinsfile` — see above.
- Fwiw, webhooks are turned off at the moment, and the build is triggered manually.
Builds, for now, are run directly on the Jenkins master as opposed to on distributed agents or within Docker containers.
### Log Fragments and crash.log files
See the [gist](https://gist.github.com/Peleke/daea9cb95c4003f6440a35feb095b9c5) for full log output.
### Additional Details / Some Questions
The error message I get is:
```
Error uploading box! Will retry in 10 seconds. Error: Put https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ: EOF[0m
```
...Which is thrown from [packer/post-processor/vagrant-cloud/step_upload.go](https://github.com/hashicorp/packer/blob/dcd21d37da053ccb2fc194c77e710ab6eb5967bf/post-processor/vagrant-cloud/step_upload.go#L17). Unfortunately, I get no additional details regarding the nature of the error:
```
2019/11/05 16:57:15 packer: 2019/11/05 16:57:15 Post-Processor Vagrant Cloud API Upload: output-vagrant/package.box https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ
2019/11/05 16:58:25 packer: 2019/11/05 16:58:25 Post-Processor Vagrant Cloud Upload Response:
2019/11/05 16:58:25 packer:
2019/11/05 16:58:25 packer: <nil>
```
My best guess (and forgive me if it's either obvious or a boneheaded hunch) is that Vagrant Cloud expects an upload of a given size but reads an EOF prior to receiving that many bytes.
I noticed these seemingly related, as-yet unresolved issues:
- [#3378](https://github.com/hashicorp/packer/issues/3378)
- #8229
If this is just my mistake, please let me know what I'm doing wrong. But if it's actually a bug and someone can point me in the right direction as to the nature of the error, I'd be happy to discuss contributing a fix.
Thanks for the time/attention :boom:
|
1.0
|
Vagrant Cloud post-processor upload errors with unexpected EOF - #### Overview of the Issue
tl;dr: Vagrant Cloud post-processor fails with an unexpected EOF when uploading a successful build targeting Virtual Box. The build is running on a Jenkins agent. I'm running the build locally at the moment, and will report back with results when it finishes.
I'm using a near-minimal Packer template with the Vagrant builder to run an Ansible build on the `bento/ubuntu-18.04` base box. Template is attached below. I've stripped it down to the bare essentials and reduced the Ansible provisioning to a series of package installs with `apt` to attempt to pinpoint the error, so there aren't many distractions.
#### Reproduction Steps
Trigger the build on a Jenkins agent via Blue Ocean pipelines.
See infrastructure details below for details.
### Packer version
Packer version `1.4.4`.
### Simplified Packer Buildfile
<details>
<summary><b>Packer template</b> (`UbuntuVM.json`)</summary>
<pre>
{
"variables": {
"cloud_token": "{{ env `VAGRANT_CLOUD_TOKEN` }}",
"version": "1.0.{{timestamp}}"
},
"builders": [{
"type": "vagrant",
"communicator": "ssh",
"provider": "virtualbox",
"source_path": "bento/ubuntu-18.04",
"add_force": true
}],
"provisioners": [{
"type": "ansible",
"playbook_file": "../provisioners/main.yml"
}],
"post-processors": [
{
"type": "vagrant-cloud",
"box_tag": "cybersecurity/UbuntuVM",
"access_token": "{{user `VAGRANT_CLOUD_TOKEN`}}",
"version": "{{user `version`}}"
}
]
}
</pre>
</details>
<details>
<summary><b>Jenkinsfile</b></summary>
<pre>
pipeline {
agent any
stages {
stage('Build') {
steps {
sh(label: 'Check directory', script: 'pwd && whoami')
sh(label: 'Verify Packer is on $PATH', script: '/usr/local/bin/packer --version')
}
}
stage('Packer') {
steps {
echo 'Packing...'
sh(label: 'Packer Build Stage!', script: '/usr/local/bin/packer build -debug UbuntuVM/packer/UbuntuVM.json')
}
}
stage('Deploy') {
steps {
echo 'Deploying...'
}
}
}
environment {
VAGRANT_CLOUD_TOKEN = 'Omitted Cause...Secrets'
}
}
</pre>
</details>
### Operating system and Environment details
**Infrastructure**
- Jenkins v2.203 running on Ubuntu 18.04, deployed to an Azure `Standard_D4S_v3` instance
- The Ubuntu instance is unmodified _except_ for the following installations:
- Vagrant v2.0.2
- Ansible v2.9.0
- Packer v1.4.4
**Build Details**
- The build is described in a declarative `Jenkinsfile` — see above.
- Fwiw, webhooks are turned off at the moment, and the build is triggered manually.
Builds, for now, are run directly on the Jenkins master as opposed to on distributed agents or within Docker containers.
### Log Fragments and crash.log files
See the [gist](https://gist.github.com/Peleke/daea9cb95c4003f6440a35feb095b9c5) for full log output.
### Additional Details / Some Questions
The error message I get is:
```
Error uploading box! Will retry in 10 seconds. Error: Put https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ: EOF[0m
```
...Which is thrown from [packer/post-processor/vagrant-cloud/step_upload.go](https://github.com/hashicorp/packer/blob/dcd21d37da053ccb2fc194c77e710ab6eb5967bf/post-processor/vagrant-cloud/step_upload.go#L17). Unfortunately, I get no additional details regarding the nature of the error:
```
2019/11/05 16:57:15 packer: 2019/11/05 16:57:15 Post-Processor Vagrant Cloud API Upload: output-vagrant/package.box https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ
2019/11/05 16:58:25 packer: 2019/11/05 16:58:25 Post-Processor Vagrant Cloud Upload Response:
2019/11/05 16:58:25 packer:
2019/11/05 16:58:25 packer: <nil>
```
My best guess (and forgive me if it's either obvious or a boneheaded hunch) is that Vagrant Cloud expects an upload of a given size but reads an EOF prior to receiving that many bytes.
I noticed these seemingly related, as-yet unresolved issues:
- [#3378](https://github.com/hashicorp/packer/issues/3378)
- #8229
If this is just my mistake, please let me know what I'm doing wrong. But if it's actually a bug and someone can point me in the right direction as to the nature of the error, I'd be happy to discuss contributing a fix.
Thanks for the time/attention :boom:
|
process
|
vagrant cloud post processor upload errors with unexpected eof overview of the issue tl dr vagrant cloud post processor fails with an unexpected eof when uploading a successful build targeting virtual box the build is running on a jenkins agent i m running the build locally at the moment and will report back with results when it finishes i m using a near minimal packer template with the vagrant builder to run an ansible build on the bento ubuntu base box template is attached below i ve stripped it down to the bare essentials and reduced the ansible provisioning to a series of package installs with apt to attempt to pinpoint the error so there aren t many distractions reproduction steps trigger the build on a jenkins agent via blue ocean pipelines see infrastructure details below for details packer version packer version simplified packer buildfile packer template ubuntuvm json variables cloud token env vagrant cloud token version timestamp builders type vagrant communicator ssh provider virtualbox source path bento ubuntu add force true provisioners type ansible playbook file provisioners main yml post processors type vagrant cloud box tag cybersecurity ubuntuvm access token user vagrant cloud token version user version jenkinsfile pipeline agent any stages stage build steps sh label check directory script pwd whoami sh label verify packer is on path script usr local bin packer version stage packer steps echo packing sh label packer build stage script usr local bin packer build debug ubuntuvm packer ubuntuvm json stage deploy steps echo deploying environment vagrant cloud token omitted cause secrets operating system and environment details infrastructure jenkins running on ubuntu deployed to an azure standard instance the ubuntu instance is unmodified except for the following installations vagrant ansible packer build details the build is described in a declarative jenkinsfile — see above fwiw webhooks are turned off at the moment and the build is triggered manually builds for now are run directly on the jenkins master as opposed to on distributed agents or within docker containers log fragments and crash log files see the for full log output additional details some questions the error message i get is error uploading box will retry in seconds error put eof which is thrown from unfortunately i get no additional details regarding the nature of the error packer post processor vagrant cloud api upload output vagrant package box packer post processor vagrant cloud upload response packer packer my best guess and forgive me if it s either obvious or a boneheaded hunch is that vagrant cloud expects an upload of a given size but reads an eof prior to receiving that many bytes i noticed these seemingly related as yet unresolved issues if this is just my mistake please let me know what i m doing wrong but if it s actually a bug and someone can point me in the right direction as to the nature of the error i d be happy to discuss contributing a fix thanks for the time attention boom
| 1
|
5,950
| 8,775,089,000
|
IssuesEvent
|
2018-12-18 21:52:49
|
rubberduck-vba/Rubberduck
|
https://api.github.com/repos/rubberduck-vba/Rubberduck
|
closed
|
Static Analysis & Code Metrics
|
difficulty-04-quackhead enhancement parse-tree-processing up-for-grabs user-interface
|
Would be cool to support metrics and graphs over time. Can start with "simple" metrics like Lines of Code, Cyclomatic Complexity, Maintainability Index, number of broken rules, etc.
|
1.0
|
Static Analysis & Code Metrics - Would be cool to support metrics and graphs over time. Can start with "simple" metrics like Lines of Code, Cyclomatic Complexity, Maintainability Index, number of broken rules, etc.
|
process
|
static analysis code metrics would be cool to support metrics and graphs over time can start with simple metrics like lines of code cyclomatic complexity maintainability index number of broken rules etc
| 1
|
20,816
| 27,578,247,130
|
IssuesEvent
|
2023-03-08 14:31:53
|
ukri-excalibur/excalibur-tests
|
https://api.github.com/repos/ukri-excalibur/excalibur-tests
|
opened
|
Create high-level script to run postprocessing
|
UCL postprocessing
|
This should accept some input either in command line or a config yaml/json file, and based on that it should perform the required type of analysis (see use cases in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139).
The steps of this analysis should be something like
- read desired data from some file path and store in a pandas dataframe
- produce required output (plots and/or tables)
|
1.0
|
Create high-level script to run postprocessing - This should accept some input either in command line or a config yaml/json file, and based on that it should perform the required type of analysis (see use cases in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139).
The steps of this analysis should be something like
- read desired data from some file path and store in a pandas dataframe
- produce required output (plots and/or tables)
|
process
|
create high level script to run postprocessing this should accept some input either in command line or a config yaml json file and based on that it should perform the required type of analysis see use cases in the steps of this analysis should be something like read desired data from some file path and store in a pandas dataframe produce required output plots and or tables
| 1
|
123,713
| 12,216,324,599
|
IssuesEvent
|
2020-05-01 14:54:57
|
monoDriveIO/documentation
|
https://api.github.com/repos/monoDriveIO/documentation
|
opened
|
Create generic API document
|
documentation enhancement
|
Create a generic API introduction that discusses how to send/receive messages to the simulator. This should focus on emphasizing that only messaging is required to actually control the simulation in any language.
|
1.0
|
Create generic API document - Create a generic API introduction that discusses how to send/receive messages to the simulator. This should focus on emphasizing that only messaging is required to actually control the simulation in any language.
|
non_process
|
create generic api document create a generic api introduction that discusses how to send receive messages to the simulator this should focus on emphasizing that only messaging is required to actually control the simulation in any language
| 0
|
14,553
| 17,670,747,059
|
IssuesEvent
|
2021-08-23 05:35:13
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[feature] add Fill noData cells algorithm and tests
|
Automatic new feature Processing Alg 3.14
|
Original commit: https://github.com/qgis/QGIS/commit/ea7167e135b47174b240ae4306406a7b3ee8189f by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
1.0
|
[feature] add Fill noData cells algorithm and tests - Original commit: https://github.com/qgis/QGIS/commit/ea7167e135b47174b240ae4306406a7b3ee8189f by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
process
|
add fill nodata cells algorithm and tests original commit by nyalldawson unfortunately this naughty coder did not write a description
| 1
|
250,554
| 7,978,703,077
|
IssuesEvent
|
2018-07-17 19:11:32
|
lupino3/edumips64
|
https://api.github.com/repos/lupino3/edumips64
|
opened
|
Get rid of JavaHelp in favor of JavaFX WebView
|
priority:1 type:enhancement
|
Since Java 8, the JRE includes a [WebView component](https://docs.oracle.com/javase/8/javafx/api/javafx/scene/web/WebView.html), which should be easy to integrate in a standard Swing frame (see [JavaFX for Swing](https://docs.oracle.com/javafx/2/swing/swing-fx-interoperability.htm)).
Since EduMIPS64 already depends on Java 8+, it would be great to get rid of JavaHelp, which is a dying technology born 10+ years ago and which is as of today the only external dependency for the simulator.
|
1.0
|
Get rid of JavaHelp in favor of JavaFX WebView - Since Java 8, the JRE includes a [WebView component](https://docs.oracle.com/javase/8/javafx/api/javafx/scene/web/WebView.html), which should be easy to integrate in a standard Swing frame (see [JavaFX for Swing](https://docs.oracle.com/javafx/2/swing/swing-fx-interoperability.htm)).
Since EduMIPS64 already depends on Java 8+, it would be great to get rid of JavaHelp, which is a dying technology born 10+ years ago and which is as of today the only external dependency for the simulator.
|
non_process
|
get rid of javahelp in favor of javafx webview since java the jre includes a which should be easy to integrate in a standard swing frame see since already depends on java it would be great to get rid of javahelp which is a dying technology born years ago and which is as of today the only external dependency for the simulator
| 0
|
167,807
| 20,726,409,066
|
IssuesEvent
|
2022-03-14 02:49:30
|
dgee2/dgee2.github.io
|
https://api.github.com/repos/dgee2/dgee2.github.io
|
opened
|
CVE-2021-33502 (High) detected in normalize-url-4.5.0.tgz
|
security vulnerability
|
## CVE-2021-33502 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-4.5.0.tgz</b></p></summary>
<p>Normalize a URL</p>
<p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gatsby-plugin-sharp/node_modules/normalize-url/package.json,/node_modules/package-json/node_modules/normalize-url/package.json,/node_modules/gatsby-source-filesystem/node_modules/normalize-url/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-plugin-sharp-3.2.0.tgz (Root Library)
- got-10.7.0.tgz
- cacheable-request-7.0.1.tgz
- :x: **normalize-url-4.5.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs.
<p>Publish Date: 2021-05-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p>
<p>Release Date: 2021-05-24</p>
<p>Fix Resolution (normalize-url): 4.5.1</p>
<p>Direct dependency fix Resolution (gatsby-plugin-sharp): 3.2.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-33502 (High) detected in normalize-url-4.5.0.tgz - ## CVE-2021-33502 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-4.5.0.tgz</b></p></summary>
<p>Normalize a URL</p>
<p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gatsby-plugin-sharp/node_modules/normalize-url/package.json,/node_modules/package-json/node_modules/normalize-url/package.json,/node_modules/gatsby-source-filesystem/node_modules/normalize-url/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-plugin-sharp-3.2.0.tgz (Root Library)
- got-10.7.0.tgz
- cacheable-request-7.0.1.tgz
- :x: **normalize-url-4.5.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs.
<p>Publish Date: 2021-05-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p>
<p>Release Date: 2021-05-24</p>
<p>Fix Resolution (normalize-url): 4.5.1</p>
<p>Direct dependency fix Resolution (gatsby-plugin-sharp): 3.2.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in normalize url tgz cve high severity vulnerability vulnerable library normalize url tgz normalize a url library home page a href path to dependency file package json path to vulnerable library node modules gatsby plugin sharp node modules normalize url package json node modules package json node modules normalize url package json node modules gatsby source filesystem node modules normalize url package json dependency hierarchy gatsby plugin sharp tgz root library got tgz cacheable request tgz x normalize url tgz vulnerable library found in base branch master vulnerability details the normalize url package before x before and x before for node js has a redos regular expression denial of service issue because it has exponential performance for data urls publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution normalize url direct dependency fix resolution gatsby plugin sharp step up your open source security game with whitesource
| 0
|
122,856
| 12,167,373,648
|
IssuesEvent
|
2020-04-27 10:48:16
|
kubeflow/tf-operator
|
https://api.github.com/repos/kubeflow/tf-operator
|
closed
|
Update instructions and code to work with Kubernetes 1.8
|
kind/documentation lifecycle/stale
|
With Kubernetes 1.8
* GPU type has changed and is now
```
nvidia.com/gpu
```
* Volume mounting NVIDIA drivers from the host should no longer be necessary if the cluster use [device plugins](https://kubernetes.io/docs/concepts/cluster-administration/device-plugins/)
@sozercan @wbuchwalter Is Azure using device plugins for GPUs or do you still rely on volume mounting GPU drivers?
|
1.0
|
Update instructions and code to work with Kubernetes 1.8 - With Kubernetes 1.8
* GPU type has changed and is now
```
nvidia.com/gpu
```
* Volume mounting NVIDIA drivers from the host should no longer be necessary if the cluster use [device plugins](https://kubernetes.io/docs/concepts/cluster-administration/device-plugins/)
@sozercan @wbuchwalter Is Azure using device plugins for GPUs or do you still rely on volume mounting GPU drivers?
|
non_process
|
update instructions and code to work with kubernetes with kubernetes gpu type has changed and is now nvidia com gpu volume mounting nvidia drivers from the host should no longer be necessary if the cluster use sozercan wbuchwalter is azure using device plugins for gpus or do you still rely on volume mounting gpu drivers
| 0
|
7,311
| 10,449,982,029
|
IssuesEvent
|
2019-09-19 09:33:50
|
bisq-network/bisq
|
https://api.github.com/repos/bisq-network/bisq
|
closed
|
Add chat system for traders communcation
|
in:gui in:trade-process
|
Once the trade has started the peers should have the ability to communicate with each other via a Bisq internal chat system (same as used in arbitration).
It requires clear rules and the ability to ignore communication if user prefers.
Any violation of the rules (e.g. trying social engineering scam) will lead to a dispute where the violating peer risks to lose his security deposit.
With the messaging system in place it would be handy if the buyer could release the btc to the seller also, in case the buyer is unable to pay for some reason.
|
1.0
|
Add chat system for traders communcation - Once the trade has started the peers should have the ability to communicate with each other via a Bisq internal chat system (same as used in arbitration).
It requires clear rules and the ability to ignore communication if user prefers.
Any violation of the rules (e.g. trying social engineering scam) will lead to a dispute where the violating peer risks to lose his security deposit.
With the messaging system in place it would be handy if the buyer could release the btc to the seller also, in case the buyer is unable to pay for some reason.
|
process
|
add chat system for traders communcation once the trade has started the peers should have the ability to communicate with each other via a bisq internal chat system same as used in arbitration it requires clear rules and the ability to ignore communication if user prefers any violation of the rules e g trying social engineering scam will lead to a dispute where the violating peer risks to lose his security deposit with the messaging system in place it would be handy if the buyer could release the btc to the seller also in case the buyer is unable to pay for some reason
| 1
|
9,451
| 12,429,236,838
|
IssuesEvent
|
2020-05-25 08:04:21
|
varys-main/ps-tools
|
https://api.github.com/repos/varys-main/ps-tools
|
opened
|
Docker - Anpassungen
|
processing
|
# User Story
- Die Rechtschreibung soll angepasst werden und die Antworten auf Abfragen (J/N) sollen standardisiert werden. Zudem soll die Standard-Option besser ersichtlich sein.
- Das Laden des Moduls NAV-ContainerHelper soll besser erklärt werden und/oder in eine eigene Funktion ausgelagert werden. Das Modul soll aktualisiert werden können (keine Installation mehrerer Versionen)
- Die gu-Module sollen bereinigt werden.
# Tasks
- [ ] Rechtschreibung/Erklärung
- [ ] NAV-ContainerHelper
- [ ] gu-Module
# Implementation
# Known Problems
|
1.0
|
Docker - Anpassungen - # User Story
- Die Rechtschreibung soll angepasst werden und die Antworten auf Abfragen (J/N) sollen standardisiert werden. Zudem soll die Standard-Option besser ersichtlich sein.
- Das Laden des Moduls NAV-ContainerHelper soll besser erklärt werden und/oder in eine eigene Funktion ausgelagert werden. Das Modul soll aktualisiert werden können (keine Installation mehrerer Versionen)
- Die gu-Module sollen bereinigt werden.
# Tasks
- [ ] Rechtschreibung/Erklärung
- [ ] NAV-ContainerHelper
- [ ] gu-Module
# Implementation
# Known Problems
|
process
|
docker anpassungen user story die rechtschreibung soll angepasst werden und die antworten auf abfragen j n sollen standardisiert werden zudem soll die standard option besser ersichtlich sein das laden des moduls nav containerhelper soll besser erklärt werden und oder in eine eigene funktion ausgelagert werden das modul soll aktualisiert werden können keine installation mehrerer versionen die gu module sollen bereinigt werden tasks rechtschreibung erklärung nav containerhelper gu module implementation known problems
| 1
|
722,124
| 24,850,850,600
|
IssuesEvent
|
2022-10-26 19:55:50
|
bounswe/bounswe2022group8
|
https://api.github.com/repos/bounswe/bounswe2022group8
|
opened
|
FE-4: Connection Between Backend and Frontend
|
Effort: High Priority: High Status: review needed coding frontend
|
### What's up?
With the completion of backend part of login and signup, we have to connect backend to frontend by providing api calls on frontend and solving some authorization issues on frontend. To do it, first I have to configure backend files and database connection to backend on my local to test if connection is done successfully.
### To Do
- [ ] Configure backend tools on local to test if api call works
- [ ] Connect signup to backend by providing api calls on frontend
- [ ] Connect login to backend by providing api calls on frontend
### Deadline
23.10.2022 @12.00
### Additional Information
_No response_
### Reviewers
@kostanya @mumcusena
|
1.0
|
FE-4: Connection Between Backend and Frontend - ### What's up?
With the completion of backend part of login and signup, we have to connect backend to frontend by providing api calls on frontend and solving some authorization issues on frontend. To do it, first I have to configure backend files and database connection to backend on my local to test if connection is done successfully.
### To Do
- [ ] Configure backend tools on local to test if api call works
- [ ] Connect signup to backend by providing api calls on frontend
- [ ] Connect login to backend by providing api calls on frontend
### Deadline
23.10.2022 @12.00
### Additional Information
_No response_
### Reviewers
@kostanya @mumcusena
|
non_process
|
fe connection between backend and frontend what s up with the completion of backend part of login and signup we have to connect backend to frontend by providing api calls on frontend and solving some authorization issues on frontend to do it first i have to configure backend files and database connection to backend on my local to test if connection is done successfully to do configure backend tools on local to test if api call works connect signup to backend by providing api calls on frontend connect login to backend by providing api calls on frontend deadline additional information no response reviewers kostanya mumcusena
| 0
|
15,781
| 27,884,245,655
|
IssuesEvent
|
2023-03-21 22:06:52
|
adrrf/acme-life-long-learning
|
https://api.github.com/repos/adrrf/acme-life-long-learning
|
opened
|
d3-s3-fr-16 - assistant operations on dashboards
|
develop functional_requirement
|
Operations by assistants on assistant dashboards:
- [ ] Show their assistant dashboards.
|
1.0
|
d3-s3-fr-16 - assistant operations on dashboards - Operations by assistants on assistant dashboards:
- [ ] Show their assistant dashboards.
|
non_process
|
fr assistant operations on dashboards operations by assistants on assistant dashboards show their assistant dashboards
| 0
|
648,354
| 21,183,748,402
|
IssuesEvent
|
2022-04-08 10:29:53
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
reopened
|
run.jobs.e2e_test: test_end_to_end failed
|
priority: p1 type: bug api: run samples flakybot: issue flakybot: flaky
|
Note: #7511 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: fe8dd4ce6953f40afa7e6b5b797b22e7f4daf546
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/fbd7f62e-c5f5-4259-b0df-d003ef8c6303), [Sponge](http://sponge2/fbd7f62e-c5f5-4259-b0df-d003ef8c6303)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/run/jobs/e2e_test.py", line 38, in setup_job
subprocess.check_call(
File "/usr/local/lib/python3.9/subprocess.py", line 373, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['gcloud', 'builds', 'submit', '--config', 'e2e_test_setup.yaml', '--project', 'python-docs-samples-tests', '--substitutions', '_SERVICE=job-quickstart-0cab165ab6,_VERSION=0cab165ab6']' returned non-zero exit status 1.</pre></details>
|
1.0
|
run.jobs.e2e_test: test_end_to_end failed - Note: #7511 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: fe8dd4ce6953f40afa7e6b5b797b22e7f4daf546
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/fbd7f62e-c5f5-4259-b0df-d003ef8c6303), [Sponge](http://sponge2/fbd7f62e-c5f5-4259-b0df-d003ef8c6303)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/run/jobs/e2e_test.py", line 38, in setup_job
subprocess.check_call(
File "/usr/local/lib/python3.9/subprocess.py", line 373, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['gcloud', 'builds', 'submit', '--config', 'e2e_test_setup.yaml', '--project', 'python-docs-samples-tests', '--substitutions', '_SERVICE=job-quickstart-0cab165ab6,_VERSION=0cab165ab6']' returned non-zero exit status 1.</pre></details>
|
non_process
|
run jobs test test end to end failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace run jobs test py line in setup job subprocess check call file usr local lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status
| 0
|
1,947
| 4,770,557,438
|
IssuesEvent
|
2016-10-26 15:33:56
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
closed
|
Rename url column to avoid conflicts in payload
|
API Explorer Processors
|
Currently `documents`, `files` and `sources` tables have a `url` column. This conflicts with our convention to send the OpenTrials API URL in the payload.
Ex:
```
{
total_count: 100,
items: [
{
id: 'document_id',
url: 'http://api.opentrials.net/v1/documents/document_id',
name: 'name',
// ...
} // ...
```
So we should rename the `url` column to `source_url` for `documents`, `files` and `sources`.
|
1.0
|
Rename url column to avoid conflicts in payload - Currently `documents`, `files` and `sources` tables have a `url` column. This conflicts with our convention to send the OpenTrials API URL in the payload.
Ex:
```
{
total_count: 100,
items: [
{
id: 'document_id',
url: 'http://api.opentrials.net/v1/documents/document_id',
name: 'name',
// ...
} // ...
```
So we should rename the `url` column to `source_url` for `documents`, `files` and `sources`.
|
process
|
rename url column to avoid conflicts in payload currently documents files and sources tables have a url column this conflicts with our convention to send the opentrials api url in the payload ex total count items id document id url name name so we should rename the url column to source url for documents files and sources
| 1
|
301,650
| 9,222,914,459
|
IssuesEvent
|
2019-03-12 01:01:27
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
opened
|
secure.meetup.com - design is broken
|
browser-firefox priority-important
|
<!-- @browser: Firefox 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:66.0) Gecko/20100101 Firefox/66.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://secure.meetup.com/login/
**Browser / Version**: Firefox 66.0
**Operating System**: Mac OS X 10.14
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Login with facebook won't work with firefox dev edition, works fine with google chrome
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190307095232</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-T2LNGD&l=gtmDataLayer was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://js-agent.newrelic.com/nr-spa-1118.min.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
secure.meetup.com - design is broken - <!-- @browser: Firefox 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:66.0) Gecko/20100101 Firefox/66.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://secure.meetup.com/login/
**Browser / Version**: Firefox 66.0
**Operating System**: Mac OS X 10.14
**Tested Another Browser**: Yes
**Problem type**: Design is broken
**Description**: Login with facebook won't work with firefox dev edition, works fine with google chrome
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190307095232</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-T2LNGD&l=gtmDataLayer was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://js-agent.newrelic.com/nr-spa-1118.min.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
secure meetup com design is broken url browser version firefox operating system mac os x tested another browser yes problem type design is broken description login with facebook won t work with firefox dev edition works fine with google chrome steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel aurora console messages u u u u u u u u u u u u u u u u u u from with ❤️
| 0
|
152,154
| 12,093,269,905
|
IssuesEvent
|
2020-04-19 18:57:00
|
gisellemartel/CONPASS
|
https://api.github.com/repos/gisellemartel/CONPASS
|
opened
|
AT-25 - (US5B - As a user, I want to get map directions to external points of interest.)
|
Acceptance Test (SPRINT 4)
|
1. Launch application
2. Select a building that supports indoor mode such as Hall and enter indoor mode by clicking on the building
3. Click on the direction button on the bottom right hand corner (blue button)
4. Enter a starting point (as you are inside)
5. Enter a destination outside like a local Tim Hortons
|
1.0
|
AT-25 - (US5B - As a user, I want to get map directions to external points of interest.) - 1. Launch application
2. Select a building that supports indoor mode such as Hall and enter indoor mode by clicking on the building
3. Click on the direction button on the bottom right hand corner (blue button)
4. Enter a starting point (as you are inside)
5. Enter a destination outside like a local Tim Hortons
|
non_process
|
at as a user i want to get map directions to external points of interest launch application select a building that supports indoor mode such as hall and enter indoor mode by clicking on the building click on the direction button on the bottom right hand corner blue button enter a starting point as you are inside enter a destination outside like a local tim hortons
| 0
|
16,693
| 21,792,169,331
|
IssuesEvent
|
2022-05-15 04:02:20
|
amor71/LiuAlgoTrader
|
https://api.github.com/repos/amor71/LiuAlgoTrader
|
closed
|
use contextvars
|
in-process no-issue-activity
|
Use python `contextvars` similar to how FastApi uses them to pass data to strategies and internally to simplify strategy development and code improvements.
Need to POC to check the impact on performance
|
1.0
|
use contextvars - Use python `contextvars` similar to how FastApi uses them to pass data to strategies and internally to simplify strategy development and code improvements.
Need to POC to check the impact on performance
|
process
|
use contextvars use python contextvars similar to how fastapi uses them to pass data to strategies and internally to simplify strategy development and code improvements need to poc to check the impact on performance
| 1
|
427,640
| 29,831,641,300
|
IssuesEvent
|
2023-06-18 10:52:09
|
io7m/medrina
|
https://api.github.com/repos/io7m/medrina
|
opened
|
Consider adjusting the syntax
|
documentation enhancement
|
Currently, rules look like this:
```
[allow-immediately
[subject [with-any-roles inventory.admin]]
[object true]
[action true]
]
```
That's fine, but the problem it creates is that nothing actually says "this is a rule" explicitly. This means that it's hard to add new types of declarations later. This might be better:
```
[rule
[allow-immediately
[subject [with-any-roles inventory.admin]]
[object true]
[action true]
]]
```
This might also interact with #3 to yield something like:
```
[rule
[description "Administrators are allowed to do anything."]
[allow-immediately
[subject [with-any-roles inventory.admin]]
[object true]
[action true]
]]
```
|
1.0
|
Consider adjusting the syntax - Currently, rules look like this:
```
[allow-immediately
[subject [with-any-roles inventory.admin]]
[object true]
[action true]
]
```
That's fine, but the problem it creates is that nothing actually says "this is a rule" explicitly. This means that it's hard to add new types of declarations later. This might be better:
```
[rule
[allow-immediately
[subject [with-any-roles inventory.admin]]
[object true]
[action true]
]]
```
This might also interact with #3 to yield something like:
```
[rule
[description "Administrators are allowed to do anything."]
[allow-immediately
[subject [with-any-roles inventory.admin]]
[object true]
[action true]
]]
```
|
non_process
|
consider adjusting the syntax currently rules look like this allow immediately that s fine but the problem it creates is that nothing actually says this is a rule explicitly this means that it s hard to add new types of declarations later this might be better rule allow immediately this might also interact with to yield something like rule allow immediately
| 0
|
114,228
| 9,693,445,938
|
IssuesEvent
|
2019-05-24 16:05:40
|
rubenverhoef/OAP-Config
|
https://api.github.com/repos/rubenverhoef/OAP-Config
|
opened
|
[HW] old screen cannot enable relay if it is on, only if it is off
|
Hardware HighPrio Needs Testing in Car bug
|
When the radio is on, it cannot enable the relay. I think the output which is used right now from the old screen is too weak.
Related #45
|
1.0
|
[HW] old screen cannot enable relay if it is on, only if it is off - When the radio is on, it cannot enable the relay. I think the output which is used right now from the old screen is too weak.
Related #45
|
non_process
|
old screen cannot enable relay if it is on only if it is off when the radio is on it cannot enable the relay i think the output which is used right now from the old screen is too weak related
| 0
|
139,210
| 31,280,660,641
|
IssuesEvent
|
2023-08-22 09:22:35
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Add an additional MatcherTransformType.CAMEL_KEEP_INTERNAL_CASING and PASCAL_KEEP_INTERNAL_CASING
|
T: Enhancement C: Code Generation P: Medium E: All Editions
|
The use case is:
> For "passwordEncrypted", I want the generated getter to be "getPasswordEncrypted()".
This is currently not possible. We would need a new `MatcherTransformType`, which transforms `a_bbbCd` into `aBbbCd` (`CAMEL_KEEP_INTERNAL_CASING`) or `ABbbCd` (`PASCAL_KEEP_INTERNAL_CASING`)
---
See also this thread here:
https://groups.google.com/d/msg/jooq-user/1iy0EdWe_T8/l7e3o_KTt-QJ
|
1.0
|
Add an additional MatcherTransformType.CAMEL_KEEP_INTERNAL_CASING and PASCAL_KEEP_INTERNAL_CASING - The use case is:
> For "passwordEncrypted", I want the generated getter to be "getPasswordEncrypted()".
This is currently not possible. We would need a new `MatcherTransformType`, which transforms `a_bbbCd` into `aBbbCd` (`CAMEL_KEEP_INTERNAL_CASING`) or `ABbbCd` (`PASCAL_KEEP_INTERNAL_CASING`)
---
See also this thread here:
https://groups.google.com/d/msg/jooq-user/1iy0EdWe_T8/l7e3o_KTt-QJ
|
non_process
|
add an additional matchertransformtype camel keep internal casing and pascal keep internal casing the use case is for passwordencrypted i want the generated getter to be getpasswordencrypted this is currently not possible we would need a new matchertransformtype which transforms a bbbcd into abbbcd camel keep internal casing or abbbcd pascal keep internal casing see also this thread here
| 0
|
17,709
| 23,602,549,975
|
IssuesEvent
|
2022-08-24 04:29:28
|
arcus-azure/arcus.messaging
|
https://api.github.com/repos/arcus-azure/arcus.messaging
|
closed
|
Override general `IMessageRouter` methods to include Azure Service Bus request tracking
|
integration:service-bus area:message-processing telemetry
|
**Is your feature request related to a problem? Please describe.**
Currently, we don't override the general message routing methods of the `IMessageRouter` interface (delegated via the general `MessageRouter`), whicih means that call via the general message routing methods won't result in Azure Service Bus telemetry requests.
**Describe the solution you'd like**
Override the general `RouteMessageAsync` method in the `AzureServiceBusMessageRouter` to include Azure Service Bus telemetry request tracking, like the other more Azure Service Bus-specific routing methods have.
|
1.0
|
Override general `IMessageRouter` methods to include Azure Service Bus request tracking - **Is your feature request related to a problem? Please describe.**
Currently, we don't override the general message routing methods of the `IMessageRouter` interface (delegated via the general `MessageRouter`), whicih means that call via the general message routing methods won't result in Azure Service Bus telemetry requests.
**Describe the solution you'd like**
Override the general `RouteMessageAsync` method in the `AzureServiceBusMessageRouter` to include Azure Service Bus telemetry request tracking, like the other more Azure Service Bus-specific routing methods have.
|
process
|
override general imessagerouter methods to include azure service bus request tracking is your feature request related to a problem please describe currently we don t override the general message routing methods of the imessagerouter interface delegated via the general messagerouter whicih means that call via the general message routing methods won t result in azure service bus telemetry requests describe the solution you d like override the general routemessageasync method in the azureservicebusmessagerouter to include azure service bus telemetry request tracking like the other more azure service bus specific routing methods have
| 1
|
9,537
| 2,615,155,841
|
IssuesEvent
|
2015-03-01 06:34:00
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
Link at "overview on offline apps" in Client-Side Storage is broken
|
auto-migrated Priority-Medium Type-Defect
|
```
Link at "overview on offline apps" in "Summary" section of
http://www.html5rocks.com/tutorials/offline/storage/ (Client-Side Storage) lead
to 404 page
I think it may link to
http://www.html5rocks.com/tutorials/offline/whats-offline/ (missing / in target
URL)
```
Original issue reported on code.google.com by `mementot...@gmail.com` on 19 Nov 2010 at 9:50
|
1.0
|
Link at "overview on offline apps" in Client-Side Storage is broken - ```
Link at "overview on offline apps" in "Summary" section of
http://www.html5rocks.com/tutorials/offline/storage/ (Client-Side Storage) lead
to 404 page
I think it may link to
http://www.html5rocks.com/tutorials/offline/whats-offline/ (missing / in target
URL)
```
Original issue reported on code.google.com by `mementot...@gmail.com` on 19 Nov 2010 at 9:50
|
non_process
|
link at overview on offline apps in client side storage is broken link at overview on offline apps in summary section of client side storage lead to page i think it may link to missing in target url original issue reported on code google com by mementot gmail com on nov at
| 0
|
59,117
| 14,529,024,068
|
IssuesEvent
|
2020-12-14 17:16:14
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
iam.api-client.access_test: test_set_policy failed
|
api: iam buildcop: issue priority: p1 samples type: bug
|
Note: #3615 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 02abb311f20f81bf7b68b4064b7ceefe0f7018ed
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cb4a3071-d944-44cb-a238-f6a2298a23b3), [Sponge](http://sponge2/cb4a3071-d944-44cb-a238-f6a2298a23b3)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/iam/api-client/access_test.py", line 91, in test_set_policy
test_call()
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 49, in wrapped_f
return Retrying(*dargs, **dkw).call(f, *args, **kw)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 206, in call
return attempt.get(self._wrap_exception)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 247, in get
six.reraise(self.value[0], self.value[1], self.value[2])
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/six.py", line 703, in reraise
raise value
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 200, in call
attempt = Attempt(fn(*args, **kwargs), attempt_number, False)
File "/workspace/iam/api-client/access_test.py", line 87, in test_call
policy = access.get_policy(GCLOUD_PROJECT, version=3)
File "/workspace/iam/api-client/access.py", line 40, in get_policy
service.projects()
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/_helpers.py", line 134, in positional_wrapper
return wrapped(*args, **kwargs)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 900, in execute
resp, content = _retry_request(
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 177, in _retry_request
resp, content = http.request(uri, method, *args, **kwargs)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google_auth_httplib2.py", line 189, in request
self.credentials.before_request(
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/auth/credentials.py", line 133, in before_request
self.refresh(request)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/service_account.py", line 361, in refresh
access_token, expiry, _ = _client.jwt_grant(request, self._token_uri, assertion)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 153, in jwt_grant
response_data = _token_endpoint_request(request, token_uri, body)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 111, in _token_endpoint_request
response_data = json.loads(response_body)
File "/usr/local/lib/python3.8/json/__init__.py", line 357, in loads
return _default_decoder.decode(s)
File "/usr/local/lib/python3.8/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/local/lib/python3.8/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)</pre></details>
|
1.0
|
iam.api-client.access_test: test_set_policy failed - Note: #3615 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 02abb311f20f81bf7b68b4064b7ceefe0f7018ed
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cb4a3071-d944-44cb-a238-f6a2298a23b3), [Sponge](http://sponge2/cb4a3071-d944-44cb-a238-f6a2298a23b3)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/iam/api-client/access_test.py", line 91, in test_set_policy
test_call()
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 49, in wrapped_f
return Retrying(*dargs, **dkw).call(f, *args, **kw)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 206, in call
return attempt.get(self._wrap_exception)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 247, in get
six.reraise(self.value[0], self.value[1], self.value[2])
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/six.py", line 703, in reraise
raise value
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 200, in call
attempt = Attempt(fn(*args, **kwargs), attempt_number, False)
File "/workspace/iam/api-client/access_test.py", line 87, in test_call
policy = access.get_policy(GCLOUD_PROJECT, version=3)
File "/workspace/iam/api-client/access.py", line 40, in get_policy
service.projects()
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/_helpers.py", line 134, in positional_wrapper
return wrapped(*args, **kwargs)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 900, in execute
resp, content = _retry_request(
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 177, in _retry_request
resp, content = http.request(uri, method, *args, **kwargs)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google_auth_httplib2.py", line 189, in request
self.credentials.before_request(
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/auth/credentials.py", line 133, in before_request
self.refresh(request)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/service_account.py", line 361, in refresh
access_token, expiry, _ = _client.jwt_grant(request, self._token_uri, assertion)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 153, in jwt_grant
response_data = _token_endpoint_request(request, token_uri, body)
File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 111, in _token_endpoint_request
response_data = json.loads(response_body)
File "/usr/local/lib/python3.8/json/__init__.py", line 357, in loads
return _default_decoder.decode(s)
File "/usr/local/lib/python3.8/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/local/lib/python3.8/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)</pre></details>
|
non_process
|
iam api client access test test set policy failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace iam api client access test py line in test set policy test call file workspace iam api client nox py lib site packages retrying py line in wrapped f return retrying dargs dkw call f args kw file workspace iam api client nox py lib site packages retrying py line in call return attempt get self wrap exception file workspace iam api client nox py lib site packages retrying py line in get six reraise self value self value self value file workspace iam api client nox py lib site packages six py line in reraise raise value file workspace iam api client nox py lib site packages retrying py line in call attempt attempt fn args kwargs attempt number false file workspace iam api client access test py line in test call policy access get policy gcloud project version file workspace iam api client access py line in get policy service projects file workspace iam api client nox py lib site packages googleapiclient helpers py line in positional wrapper return wrapped args kwargs file workspace iam api client nox py lib site packages googleapiclient http py line in execute resp content retry request file workspace iam api client nox py lib site packages googleapiclient http py line in retry request resp content http request uri method args kwargs file workspace iam api client nox py lib site packages google auth py line in request self credentials before request file workspace iam api client nox py lib site packages google auth credentials py line in before request self refresh request file workspace iam api client nox py lib site packages google service account py line in refresh access token expiry client jwt grant request self token uri assertion file workspace iam api client nox py lib site packages google client py line in jwt grant response data token endpoint request request token uri body file workspace iam api client nox py lib site packages google client py line in token endpoint request response data json loads response body file usr local lib json init py line in loads return default decoder decode s file usr local lib json decoder py line in decode obj end self raw decode s idx w s end file usr local lib json decoder py line in raw decode raise jsondecodeerror expecting value s err value from none json decoder jsondecodeerror expecting value line column char
| 0
|
66,985
| 14,814,070,712
|
IssuesEvent
|
2021-01-14 03:46:08
|
JoePep09/WebGoat
|
https://api.github.com/repos/JoePep09/WebGoat
|
opened
|
CVE-2018-20676 (Medium) detected in bootstrap-3.1.1.min.js
|
security vulnerability
|
## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.1.1.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: WebGoat/webgoat-lessons/challenge/src/main/resources/js/bootstrap.min.js,WebGoat/webgoat-container/target/classes/static/plugins/bootstrap/js/bootstrap.min.js,WebGoat/webgoat-container/src/main/resources/static/plugins/bootstrap/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.1.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/JoePep09/WebGoat/commit/cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2">cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-20676 (Medium) detected in bootstrap-3.1.1.min.js - ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.1.1.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: WebGoat/webgoat-lessons/challenge/src/main/resources/js/bootstrap.min.js,WebGoat/webgoat-container/target/classes/static/plugins/bootstrap/js/bootstrap.min.js,WebGoat/webgoat-container/src/main/resources/static/plugins/bootstrap/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.1.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/JoePep09/WebGoat/commit/cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2">cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library webgoat webgoat lessons challenge src main resources js bootstrap min js webgoat webgoat container target classes static plugins bootstrap js bootstrap min js webgoat webgoat container src main resources static plugins bootstrap js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch develop vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap step up your open source security game with whitesource
| 0
|
8,306
| 11,463,675,091
|
IssuesEvent
|
2020-02-07 16:27:12
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
Solid Fuel display/database
|
Process Heating
|

These should be showing up as 51.7%, 5.4%, 3.1%, 0%, 0.2%, 39.5%, and 0.1%. I think the tool is already accounting for them being in % format instead of fraction (hence "Difference" = 99%).
|
1.0
|
Solid Fuel display/database - 
These should be showing up as 51.7%, 5.4%, 3.1%, 0%, 0.2%, 39.5%, and 0.1%. I think the tool is already accounting for them being in % format instead of fraction (hence "Difference" = 99%).
|
process
|
solid fuel display database these should be showing up as and i think the tool is already accounting for them being in format instead of fraction hence difference
| 1
|
4,277
| 7,190,483,476
|
IssuesEvent
|
2018-02-02 17:24:13
|
brucemiller/LaTeXML
|
https://api.github.com/repos/brucemiller/LaTeXML
|
closed
|
ImageMagick 7 incompatibilities (was Rotation of an included graphics leads to error in latexmlpost
|
bug postprocessing
|
A simple example document with an included graphic like
```
\documentclass[12pt,a4paper,twoside]{book}
\usepackage{graphicx}
\begin{document}
\begin{figure}[t]
\includegraphics[angle=90,width=0.7\linewidth]{some.eps}
\end{figure}
\end{document}
```
leads to an error in the latexml post processing
```
$ latexml testgraphics.tex --dest=testgraphics.xml
...
Conversion complete: No obvious problems.
$ latexmlpost --format=html5 --dest=testgraphics.html testgraphics.xml
...
(Graphics testgraphics.html processing...
Error:imageprocessing:Rotate Image processing operation Rotate (degrees, -90, color, #FFFFFF) returned Exception 410: unrecognized option `color' @ error/Q16HDRI.xs/XS_Image__Magick__Q16HDRI_Mogrify/7807
In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic
Warning:expected:image Couldn't get usable image for some.eps
In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic
0.75 sec)
(XSLT[using LaTeXML-html5.xsl] testgraphics.html processing... 0.02 sec)
```
The message ``Exception 410: unrecognized option `color'`` seems harmless to me. Maybe the error handling is a bit too strict. I think the processing of the image should not be aborted, because of this issue.
|
1.0
|
ImageMagick 7 incompatibilities (was Rotation of an included graphics leads to error in latexmlpost - A simple example document with an included graphic like
```
\documentclass[12pt,a4paper,twoside]{book}
\usepackage{graphicx}
\begin{document}
\begin{figure}[t]
\includegraphics[angle=90,width=0.7\linewidth]{some.eps}
\end{figure}
\end{document}
```
leads to an error in the latexml post processing
```
$ latexml testgraphics.tex --dest=testgraphics.xml
...
Conversion complete: No obvious problems.
$ latexmlpost --format=html5 --dest=testgraphics.html testgraphics.xml
...
(Graphics testgraphics.html processing...
Error:imageprocessing:Rotate Image processing operation Rotate (degrees, -90, color, #FFFFFF) returned Exception 410: unrecognized option `color' @ error/Q16HDRI.xs/XS_Image__Magick__Q16HDRI_Mogrify/7807
In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic
Warning:expected:image Couldn't get usable image for some.eps
In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic
0.75 sec)
(XSLT[using LaTeXML-html5.xsl] testgraphics.html processing... 0.02 sec)
```
The message ``Exception 410: unrecognized option `color'`` seems harmless to me. Maybe the error handling is a bit too strict. I think the processing of the image should not be aborted, because of this issue.
|
process
|
imagemagick incompatibilities was rotation of an included graphics leads to error in latexmlpost a simple example document with an included graphic like documentclass book usepackage graphicx begin document begin figure includegraphics some eps end figure end document leads to an error in the latexml post processing latexml testgraphics tex dest testgraphics xml conversion complete no obvious problems latexmlpost format dest testgraphics html testgraphics xml graphics testgraphics html processing error imageprocessing rotate image processing operation rotate degrees color ffffff returned exception unrecognized option color error xs xs image magick mogrify in post graphics transformgraphic warning expected image couldn t get usable image for some eps in post graphics transformgraphic sec xslt testgraphics html processing sec the message exception unrecognized option color seems harmless to me maybe the error handling is a bit too strict i think the processing of the image should not be aborted because of this issue
| 1
|
63,275
| 26,329,935,191
|
IssuesEvent
|
2023-01-10 10:01:49
|
Azure/azure-cli
|
https://api.github.com/repos/Azure/azure-cli
|
closed
|
CognitiveServices Accounts are not able to use the list-skus feature in the az cli
|
question Cognitive Services customer-reported CXP Attention Auto-Assign issue-addressed
|
<!--- 🛑 Please check existing issues first before continuing: https://github.com/Azure/azure-cli/issues --->
### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az cognitiveservices account list-skus`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
No value for given attribute
Traceback (most recent call last):
File "/usr/lib64/az/lib/python3.9/site-packages/knack/cli.py", line 233, in invoke
cmd_result = self.invocation.execute(args)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 663, in execute
raise ex
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 726, in _run_jobs_serially
results.append(self._run_job(expanded_arg, cmd_copy))
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 697, in _run_job
result = cmd_copy(params)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 333, in __call__
return self.handler(*args, **kwargs)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/command_operation.py", line 121, in handler
return op(**command_args)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/command_modules/cognitiveservices/custom.py", line 75, in list_skus
return cf_accounts(cmd.cli_ctx).list_skus(resource_group_name, account_name)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/core/tracing/decorator.py", line 73, in wrapper_use_tracer
return func(*args, **kwargs)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 1354, in list_skus
request = build_list_skus_request(
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 330, in build_list_skus_request
"accountName": _SERIALIZER.url(
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 703, in url
output = self.serialize_data(data, data_type, **kwargs)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 780, in serialize_data
raise ValueError("No value for given attribute")
ValueError: No value for given attribute
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az cognitiveservices account list-skus --kind {}`
## Expected Behavior
## Environment Summary
```
Linux-5.4.0-1098-azure-x86_64-with-glibc2.35 (Cloud Shell), CBL-Mariner/Linux
Python 3.9.14
Installer: RPM
azure-cli 2.43.0
Extensions:
quota 0.1.0
connectedk8s 1.3.3
ai-examples 0.2.5
ml 2.11.0
ssh 1.1.3
Dependencies:
msal 1.20.0
azure-mgmt-resource 21.1.0b1
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
1.0
|
CognitiveServices Accounts are not able to use the list-skus feature in the az cli - <!--- 🛑 Please check existing issues first before continuing: https://github.com/Azure/azure-cli/issues --->
### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az cognitiveservices account list-skus`
**Errors:**
```
The command failed with an unexpected error. Here is the traceback:
No value for given attribute
Traceback (most recent call last):
File "/usr/lib64/az/lib/python3.9/site-packages/knack/cli.py", line 233, in invoke
cmd_result = self.invocation.execute(args)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 663, in execute
raise ex
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 726, in _run_jobs_serially
results.append(self._run_job(expanded_arg, cmd_copy))
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 697, in _run_job
result = cmd_copy(params)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 333, in __call__
return self.handler(*args, **kwargs)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/command_operation.py", line 121, in handler
return op(**command_args)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/command_modules/cognitiveservices/custom.py", line 75, in list_skus
return cf_accounts(cmd.cli_ctx).list_skus(resource_group_name, account_name)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/core/tracing/decorator.py", line 73, in wrapper_use_tracer
return func(*args, **kwargs)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 1354, in list_skus
request = build_list_skus_request(
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 330, in build_list_skus_request
"accountName": _SERIALIZER.url(
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 703, in url
output = self.serialize_data(data, data_type, **kwargs)
File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 780, in serialize_data
raise ValueError("No value for given attribute")
ValueError: No value for given attribute
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az cognitiveservices account list-skus --kind {}`
## Expected Behavior
## Environment Summary
```
Linux-5.4.0-1098-azure-x86_64-with-glibc2.35 (Cloud Shell), CBL-Mariner/Linux
Python 3.9.14
Installer: RPM
azure-cli 2.43.0
Extensions:
quota 0.1.0
connectedk8s 1.3.3
ai-examples 0.2.5
ml 2.11.0
ssh 1.1.3
Dependencies:
msal 1.20.0
azure-mgmt-resource 21.1.0b1
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
non_process
|
cognitiveservices accounts are not able to use the list skus feature in the az cli this is autogenerated please review and update as needed describe the bug command name az cognitiveservices account list skus errors the command failed with an unexpected error here is the traceback no value for given attribute traceback most recent call last file usr az lib site packages knack cli py line in invoke cmd result self invocation execute args file usr az lib site packages azure cli core commands init py line in execute raise ex file usr az lib site packages azure cli core commands init py line in run jobs serially results append self run job expanded arg cmd copy file usr az lib site packages azure cli core commands init py line in run job result cmd copy params file usr az lib site packages azure cli core commands init py line in call return self handler args kwargs file usr az lib site packages azure cli core commands command operation py line in handler return op command args file usr az lib site packages azure cli command modules cognitiveservices custom py line in list skus return cf accounts cmd cli ctx list skus resource group name account name file usr az lib site packages azure core tracing decorator py line in wrapper use tracer return func args kwargs file usr az lib site packages azure mgmt cognitiveservices operations accounts operations py line in list skus request build list skus request file usr az lib site packages azure mgmt cognitiveservices operations accounts operations py line in build list skus request accountname serializer url file usr az lib site packages azure mgmt cognitiveservices serialization py line in url output self serialize data data data type kwargs file usr az lib site packages azure mgmt cognitiveservices serialization py line in serialize data raise valueerror no value for given attribute valueerror no value for given attribute to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az cognitiveservices account list skus kind expected behavior environment summary linux azure with cloud shell cbl mariner linux python installer rpm azure cli extensions quota ai examples ml ssh dependencies msal azure mgmt resource additional context
| 0
|
248,899
| 7,940,735,128
|
IssuesEvent
|
2018-07-10 00:06:57
|
DLR-SC/tigl
|
https://api.github.com/repos/DLR-SC/tigl
|
closed
|
Inconsistent API CCPACSFuselage::GetLoft
|
Priority-High
|
The method CCPACSFuselage::GetLoft is inconsistent in two ways since commit fb87cd650bc39:
1. It returns a TopoDS_Shape instead of a PNamedShape
2. It has a transformation argument. The default API doen not have this argument.
This is problematic for e.g. the Python API since it overwrites the old behavior and the python examples don't work.
|
1.0
|
Inconsistent API CCPACSFuselage::GetLoft - The method CCPACSFuselage::GetLoft is inconsistent in two ways since commit fb87cd650bc39:
1. It returns a TopoDS_Shape instead of a PNamedShape
2. It has a transformation argument. The default API doen not have this argument.
This is problematic for e.g. the Python API since it overwrites the old behavior and the python examples don't work.
|
non_process
|
inconsistent api ccpacsfuselage getloft the method ccpacsfuselage getloft is inconsistent in two ways since commit it returns a topods shape instead of a pnamedshape it has a transformation argument the default api doen not have this argument this is problematic for e g the python api since it overwrites the old behavior and the python examples don t work
| 0
|
15,997
| 2,870,250,336
|
IssuesEvent
|
2015-06-07 00:34:14
|
pdelia/away3d
|
https://api.github.com/repos/pdelia/away3d
|
closed
|
Rendering problem with TextField3D and Trident
|
auto-migrated Priority-Medium Type-Defect
|
#49 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:23Z
```
What steps will reproduce the problem?
1. Add TextField3D and Trident to same Away3D scene
2. Render.
What is the expected output? What do you see instead?
In addition to the 3D text and the trident, I see additional spurious lines
and shapes.
What version of the product are you using? On what operating system?
Away3D trunk rev 1404 (17 June 2009). Flash CS3. Windows XP.
Please provide any additional information below.
FLA demonstrating the problem is attached. Same problem seems to occur if
I mix a Sprite2D with a Trident or any object using LineSegment.
```
Original issue reported on code.google.com by `ralph%sa...@gtempaccount.com` on 17 Jun 2009 at 11:20
Attachments:
* [renderBug.fla](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.fla)
* [renderBug.swf](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.swf)
|
1.0
|
Rendering problem with TextField3D and Trident - #49 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:23Z
```
What steps will reproduce the problem?
1. Add TextField3D and Trident to same Away3D scene
2. Render.
What is the expected output? What do you see instead?
In addition to the 3D text and the trident, I see additional spurious lines
and shapes.
What version of the product are you using? On what operating system?
Away3D trunk rev 1404 (17 June 2009). Flash CS3. Windows XP.
Please provide any additional information below.
FLA demonstrating the problem is attached. Same problem seems to occur if
I mix a Sprite2D with a Trident or any object using LineSegment.
```
Original issue reported on code.google.com by `ralph%sa...@gtempaccount.com` on 17 Jun 2009 at 11:20
Attachments:
* [renderBug.fla](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.fla)
* [renderBug.swf](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.swf)
|
non_process
|
rendering problem with and trident issue by googlecodeexporter created on what steps will reproduce the problem add and trident to same scene render what is the expected output what do you see instead in addition to the text and the trident i see additional spurious lines and shapes what version of the product are you using on what operating system trunk rev june flash windows xp please provide any additional information below fla demonstrating the problem is attached same problem seems to occur if i mix a with a trident or any object using linesegment original issue reported on code google com by ralph sa gtempaccount com on jun at attachments
| 0
|
89,224
| 3,790,766,636
|
IssuesEvent
|
2016-03-21 22:48:12
|
projectcalico/calico-containers
|
https://api.github.com/repos/projectcalico/calico-containers
|
closed
|
Manual setup fails to connect to bird
|
Priority: Medium
|
Running the manual, two VM setup for a trial on Centos. Docker container up, calico containers fire up and I can't seem to connect to bird socket.
$ sudo /home/cloud-user/calicoctl node --libnetwork
No IP provided. Using detected IP: 192.168.98.6
Calico node is running with id: 5277d2ec28e06e7a6bf336b9ce318a4355afd847bd648473c3f0b228cd89c169
Calico libnetwork driver is running with id: 9bc2c26970c1e91523a0ec394046d209a48d4bfe2d9bbf1ef96bbbd2fbec1427
[cloud-user@ken1 ~]$ calicoctl status
calico-node container is running. Status: Up 12 seconds
Running felix version 1.3.0rc6
IPv4 BGP status
IP: 192.168.98.6 AS Number: 64511 (inherited)
Unable to connect to server control socket (/var/run/calico/bird.ctl): Permission denied
Couldn't connect to bird. Try running as root.
IPv6 BGP status
No IPv6 address configured.
|
1.0
|
Manual setup fails to connect to bird - Running the manual, two VM setup for a trial on Centos. Docker container up, calico containers fire up and I can't seem to connect to bird socket.
$ sudo /home/cloud-user/calicoctl node --libnetwork
No IP provided. Using detected IP: 192.168.98.6
Calico node is running with id: 5277d2ec28e06e7a6bf336b9ce318a4355afd847bd648473c3f0b228cd89c169
Calico libnetwork driver is running with id: 9bc2c26970c1e91523a0ec394046d209a48d4bfe2d9bbf1ef96bbbd2fbec1427
[cloud-user@ken1 ~]$ calicoctl status
calico-node container is running. Status: Up 12 seconds
Running felix version 1.3.0rc6
IPv4 BGP status
IP: 192.168.98.6 AS Number: 64511 (inherited)
Unable to connect to server control socket (/var/run/calico/bird.ctl): Permission denied
Couldn't connect to bird. Try running as root.
IPv6 BGP status
No IPv6 address configured.
|
non_process
|
manual setup fails to connect to bird running the manual two vm setup for a trial on centos docker container up calico containers fire up and i can t seem to connect to bird socket sudo home cloud user calicoctl node libnetwork no ip provided using detected ip calico node is running with id calico libnetwork driver is running with id calicoctl status calico node container is running status up seconds running felix version bgp status ip as number inherited unable to connect to server control socket var run calico bird ctl permission denied couldn t connect to bird try running as root bgp status no address configured
| 0
|
301,648
| 26,082,932,690
|
IssuesEvent
|
2022-12-25 17:06:03
|
Sun-Mountain/lettuceMeetApp
|
https://api.github.com/repos/Sun-Mountain/lettuceMeetApp
|
closed
|
Test - Finish tests for registration and login
|
tests
|
Describe the task here and link or add dependencies to any applicable issues...
### This task is done when...
- [ ] criteria 1
- [ ] criteria 2
---
Add additional labels (design, dev, compliance, BUG, etc) and size before
submitting.
If the issue is needed to complete prioritized work for the CURRENT SPRINT, add
it to the "This Sprint" pipeline. Otherwise, all other issues will be
automatically added to the _unprioritized pipeline_ for prioritization in
backlog refinement or sprint planning with Product.
|
1.0
|
Test - Finish tests for registration and login - Describe the task here and link or add dependencies to any applicable issues...
### This task is done when...
- [ ] criteria 1
- [ ] criteria 2
---
Add additional labels (design, dev, compliance, BUG, etc) and size before
submitting.
If the issue is needed to complete prioritized work for the CURRENT SPRINT, add
it to the "This Sprint" pipeline. Otherwise, all other issues will be
automatically added to the _unprioritized pipeline_ for prioritization in
backlog refinement or sprint planning with Product.
|
non_process
|
test finish tests for registration and login describe the task here and link or add dependencies to any applicable issues this task is done when criteria criteria add additional labels design dev compliance bug etc and size before submitting if the issue is needed to complete prioritized work for the current sprint add it to the this sprint pipeline otherwise all other issues will be automatically added to the unprioritized pipeline for prioritization in backlog refinement or sprint planning with product
| 0
|
17,269
| 23,051,022,635
|
IssuesEvent
|
2022-07-24 16:24:56
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Music Hell from “We are Little Zombies” (Screenshots and Title Card Added)
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Music Hell
Type (film/tv show): TV show - entertainment commentary (music like MTV shows)
Film or show in which it appears: We Are Little Zombies
Is the parent film/show streaming anywhere? Yes - Amazon Prime
About when in the parent film/show does it appear? probably about halfway
Actual footage of the film/show can be seen (yes/no)? Yes - You can see a little bit of it in the trailer. It leads into the band, Little Zombies', music video. https://youtu.be/obstTJVi3cE?t=74
|
1.0
|
Add Music Hell from “We are Little Zombies” (Screenshots and Title Card Added) - Please add as much of the following info as you can:
Title: Music Hell
Type (film/tv show): TV show - entertainment commentary (music like MTV shows)
Film or show in which it appears: We Are Little Zombies
Is the parent film/show streaming anywhere? Yes - Amazon Prime
About when in the parent film/show does it appear? probably about halfway
Actual footage of the film/show can be seen (yes/no)? Yes - You can see a little bit of it in the trailer. It leads into the band, Little Zombies', music video. https://youtu.be/obstTJVi3cE?t=74
|
process
|
add music hell from “we are little zombies” screenshots and title card added please add as much of the following info as you can title music hell type film tv show tv show entertainment commentary music like mtv shows film or show in which it appears we are little zombies is the parent film show streaming anywhere yes amazon prime about when in the parent film show does it appear probably about halfway actual footage of the film show can be seen yes no yes you can see a little bit of it in the trailer it leads into the band little zombies music video
| 1
|
8,578
| 11,746,976,375
|
IssuesEvent
|
2020-03-12 12:49:37
|
pwittchen/InfiniteScroll
|
https://api.github.com/repos/pwittchen/InfiniteScroll
|
opened
|
release 0.0.3
|
release process
|
**Release notes**:
- updated project dependencies
- updated build configuration
**Things to do**:
- [ ] bump library version
- [ ] upload Archives to Maven Central Repository
- [ ] close and release artifact on Nexus
- [ ] update gh-pages with JavaDoc
- [ ] update `CHANGELOG.md` after Maven Sync
- [ ] update Download section in `README.md` after Maven Sync
- [ ] create GitHub release
|
1.0
|
release 0.0.3 - **Release notes**:
- updated project dependencies
- updated build configuration
**Things to do**:
- [ ] bump library version
- [ ] upload Archives to Maven Central Repository
- [ ] close and release artifact on Nexus
- [ ] update gh-pages with JavaDoc
- [ ] update `CHANGELOG.md` after Maven Sync
- [ ] update Download section in `README.md` after Maven Sync
- [ ] create GitHub release
|
process
|
release release notes updated project dependencies updated build configuration things to do bump library version upload archives to maven central repository close and release artifact on nexus update gh pages with javadoc update changelog md after maven sync update download section in readme md after maven sync create github release
| 1
|
210,073
| 7,182,970,005
|
IssuesEvent
|
2018-02-01 11:40:31
|
crowdAI/crowdai
|
https://api.github.com/repos/crowdAI/crowdai
|
opened
|
Gitlab cert
|
high priority
|
Your certificate (or certificates) for the names listed below will expire in
20 days (on 21 Feb 18 11:46 +0000). Please make sure to renew
your certificate before then, or visitors to your website will encounter errors.
gitlab.crowdai.org
|
1.0
|
Gitlab cert - Your certificate (or certificates) for the names listed below will expire in
20 days (on 21 Feb 18 11:46 +0000). Please make sure to renew
your certificate before then, or visitors to your website will encounter errors.
gitlab.crowdai.org
|
non_process
|
gitlab cert your certificate or certificates for the names listed below will expire in days on feb please make sure to renew your certificate before then or visitors to your website will encounter errors gitlab crowdai org
| 0
|
10,072
| 14,444,468,965
|
IssuesEvent
|
2020-12-07 21:17:49
|
spacetelescope/jwst
|
https://api.github.com/repos/spacetelescope/jwst
|
closed
|
exposures that get Level 2b processing
|
Requirement associations
|
Prompted by [this comment](https://github.com/STScI-JWST/jwst/issues/735#issuecomment-292226126), this issue documents exactly which exposures will go through any type of Level 2b processing, such as Spec2Pipeline or Image2Pipeline. Please review, update, and if updated, leave a comment that one has done so. Questions in following comments.
EXP_TYPE | Association Exposure Type | Specials | Level2b?
---------- | --------------------------- | --------- | ---------
NRC_IMAGE | science | n/a | YES
NRC_GRISM/NRC_WFSS | science | n/a | YES
NRC_TACQ | target_acquisition | n/a | YES
NRC_TACONFIRM | target_acquisition | n/a | YES
NRC_CORON | science | n/a | YES
NRC_CORON | psf | PSF | YES
NRC_TSIMAGE | science | n/a | YES
NRC_TSGRISM | science | n/a | YES
NRC_FOCUS | science | n/a/ | YES
NRC_DARK | dark | n/a | NO
NRC_FLAT | flat | n/a | NO
NRC_LED | science | n/a | NO
| | |
MIR_IMAGE | science | n/a | YES
MIR_TACQ | target_acquisition | n/a | YES
MIR_LYOT | science | n/a | YES
MIR_LYOT | psf | PSF | YES
MIR_4QPM | science | n/a | YES
MIR_4QPM | psf | PSF | YES
MIR_LRS-FIXEDSLIT | science | n/a | YES
MIR_LRS-FIXEDSLIT | background | BACKGROUND | YES
MIR_LRS-SLITLESS | science | n/a | YES
MIR_LRS-SLITLESS | background | BACKGROUND | YES
MIR_MRS | science | n/a | YES
MIR_MRS | background | BACKGROUND | YES
MIR_DARKIMG | dark | n/a | NO
MIR_DARKMRS | dark | n/a | NO
MIR_FLATIMAGE | flat | n/a | NO
MIR_FLATIMAGE-EXT | flat | n/a | NO
MIR_FLATMRS | flat | n/a | NO
MIR_FLATMRS-EXT | flat | n/a | NO
MIR_CORONCAL | science | n/a | YES
| | |
NRS_WATA | target_acquisition | n/a | YES
NRS_MSATA | target_acquisition | n/a | YES
NRS_TACONFIRM | target_acquisition | n/a | YES
NRS_CONFIRM | science | n/a | YES
NRS_FIXEDSLIT | science | n/a | YES
NRS_FIXEDSLIT | background | BACKGROUND | YES
NRS_AUTOWAVE | nrs_autowave | n/a | YES
NRS_IFU | science | n/a | YES
NRS_IFU | imprint | IMPRINT | YES
NRS_IFU | background | BACKGROUND | YES
NRS_IMAGE | science | n/a | YES
NRS_MSASPEC | science | n/a | YES
NRS_MSASPEC | imprint | IMPRINT | YES
NRS_AUTOFLAT | nrs_autoflat | n/a | YES
NRS_FOCUS | science | n/a | YES
NRS_DARK | dark | n/a | NO
NRS_LAMP | science | n/a | YES<sup>1</sup>
NRS_BRIGHTOBJ | science | n/a | YES
NRS_MIMF | science | n/a | YES
NRS_VERIFY | science | n/a | YES<sup>1</sup>
| | |
NIS_IMAGE | science | n/a | YES
NIS_WFSS | science | n/a | YES
NIS_TACQ | target_acquisition | n/a | YES
NIS_TACONFIRM | target_acquisition | n/a | YES
NIS_SOSS | science | n/a | YES
NIS_AMI | science | n/a | YES
NIS_AMI | psf | PSF | YES
NIS_FOCUS | science | n/a | YES
NIS_DARK | science | n/a | NO
NIS_LAMP | science | n/a | NO
NIS_EXTCAL | science | n/a | NO
| | |
FGS_IMAGE | science | n/a | YES
FGS_FOCUS | science | n/a | YES
FGS_SKYFLAT | flat | n/a | NO
FGS_INTFLAT | flat | n/a | NO
FGS_DARK | dark | n/a | NO
FGS_ID-STACK | tracking | n/a | NO
FGS_ID-IMAGE | tracking | n/a | NO
FGS_ACQ1 | tracking | n/a | NO
FGS_ACQ2 | tracking | n/a | NO
FGS_TRACK | tracking | n/a | NO
FGS_FINEGUIDE | tracking | n/a | NO
## Notes
<sup>1</sup> Custom Level2 processing
|
1.0
|
exposures that get Level 2b processing - Prompted by [this comment](https://github.com/STScI-JWST/jwst/issues/735#issuecomment-292226126), this issue documents exactly which exposures will go through any type of Level 2b processing, such as Spec2Pipeline or Image2Pipeline. Please review, update, and if updated, leave a comment that one has done so. Questions in following comments.
EXP_TYPE | Association Exposure Type | Specials | Level2b?
---------- | --------------------------- | --------- | ---------
NRC_IMAGE | science | n/a | YES
NRC_GRISM/NRC_WFSS | science | n/a | YES
NRC_TACQ | target_acquisition | n/a | YES
NRC_TACONFIRM | target_acquisition | n/a | YES
NRC_CORON | science | n/a | YES
NRC_CORON | psf | PSF | YES
NRC_TSIMAGE | science | n/a | YES
NRC_TSGRISM | science | n/a | YES
NRC_FOCUS | science | n/a/ | YES
NRC_DARK | dark | n/a | NO
NRC_FLAT | flat | n/a | NO
NRC_LED | science | n/a | NO
| | |
MIR_IMAGE | science | n/a | YES
MIR_TACQ | target_acquisition | n/a | YES
MIR_LYOT | science | n/a | YES
MIR_LYOT | psf | PSF | YES
MIR_4QPM | science | n/a | YES
MIR_4QPM | psf | PSF | YES
MIR_LRS-FIXEDSLIT | science | n/a | YES
MIR_LRS-FIXEDSLIT | background | BACKGROUND | YES
MIR_LRS-SLITLESS | science | n/a | YES
MIR_LRS-SLITLESS | background | BACKGROUND | YES
MIR_MRS | science | n/a | YES
MIR_MRS | background | BACKGROUND | YES
MIR_DARKIMG | dark | n/a | NO
MIR_DARKMRS | dark | n/a | NO
MIR_FLATIMAGE | flat | n/a | NO
MIR_FLATIMAGE-EXT | flat | n/a | NO
MIR_FLATMRS | flat | n/a | NO
MIR_FLATMRS-EXT | flat | n/a | NO
MIR_CORONCAL | science | n/a | YES
| | |
NRS_WATA | target_acquisition | n/a | YES
NRS_MSATA | target_acquisition | n/a | YES
NRS_TACONFIRM | target_acquisition | n/a | YES
NRS_CONFIRM | science | n/a | YES
NRS_FIXEDSLIT | science | n/a | YES
NRS_FIXEDSLIT | background | BACKGROUND | YES
NRS_AUTOWAVE | nrs_autowave | n/a | YES
NRS_IFU | science | n/a | YES
NRS_IFU | imprint | IMPRINT | YES
NRS_IFU | background | BACKGROUND | YES
NRS_IMAGE | science | n/a | YES
NRS_MSASPEC | science | n/a | YES
NRS_MSASPEC | imprint | IMPRINT | YES
NRS_AUTOFLAT | nrs_autoflat | n/a | YES
NRS_FOCUS | science | n/a | YES
NRS_DARK | dark | n/a | NO
NRS_LAMP | science | n/a | YES<sup>1</sup>
NRS_BRIGHTOBJ | science | n/a | YES
NRS_MIMF | science | n/a | YES
NRS_VERIFY | science | n/a | YES<sup>1</sup>
| | |
NIS_IMAGE | science | n/a | YES
NIS_WFSS | science | n/a | YES
NIS_TACQ | target_acquisition | n/a | YES
NIS_TACONFIRM | target_acquisition | n/a | YES
NIS_SOSS | science | n/a | YES
NIS_AMI | science | n/a | YES
NIS_AMI | psf | PSF | YES
NIS_FOCUS | science | n/a | YES
NIS_DARK | science | n/a | NO
NIS_LAMP | science | n/a | NO
NIS_EXTCAL | science | n/a | NO
| | |
FGS_IMAGE | science | n/a | YES
FGS_FOCUS | science | n/a | YES
FGS_SKYFLAT | flat | n/a | NO
FGS_INTFLAT | flat | n/a | NO
FGS_DARK | dark | n/a | NO
FGS_ID-STACK | tracking | n/a | NO
FGS_ID-IMAGE | tracking | n/a | NO
FGS_ACQ1 | tracking | n/a | NO
FGS_ACQ2 | tracking | n/a | NO
FGS_TRACK | tracking | n/a | NO
FGS_FINEGUIDE | tracking | n/a | NO
## Notes
<sup>1</sup> Custom Level2 processing
|
non_process
|
exposures that get level processing prompted by this issue documents exactly which exposures will go through any type of level processing such as or please review update and if updated leave a comment that one has done so questions in following comments exp type association exposure type specials nrc image science n a yes nrc grism nrc wfss science n a yes nrc tacq target acquisition n a yes nrc taconfirm target acquisition n a yes nrc coron science n a yes nrc coron psf psf yes nrc tsimage science n a yes nrc tsgrism science n a yes nrc focus science n a yes nrc dark dark n a no nrc flat flat n a no nrc led science n a no mir image science n a yes mir tacq target acquisition n a yes mir lyot science n a yes mir lyot psf psf yes mir science n a yes mir psf psf yes mir lrs fixedslit science n a yes mir lrs fixedslit background background yes mir lrs slitless science n a yes mir lrs slitless background background yes mir mrs science n a yes mir mrs background background yes mir darkimg dark n a no mir darkmrs dark n a no mir flatimage flat n a no mir flatimage ext flat n a no mir flatmrs flat n a no mir flatmrs ext flat n a no mir coroncal science n a yes nrs wata target acquisition n a yes nrs msata target acquisition n a yes nrs taconfirm target acquisition n a yes nrs confirm science n a yes nrs fixedslit science n a yes nrs fixedslit background background yes nrs autowave nrs autowave n a yes nrs ifu science n a yes nrs ifu imprint imprint yes nrs ifu background background yes nrs image science n a yes nrs msaspec science n a yes nrs msaspec imprint imprint yes nrs autoflat nrs autoflat n a yes nrs focus science n a yes nrs dark dark n a no nrs lamp science n a yes nrs brightobj science n a yes nrs mimf science n a yes nrs verify science n a yes nis image science n a yes nis wfss science n a yes nis tacq target acquisition n a yes nis taconfirm target acquisition n a yes nis soss science n a yes nis ami science n a yes nis ami psf psf yes nis focus science n a yes nis dark science n a no nis lamp science n a no nis extcal science n a no fgs image science n a yes fgs focus science n a yes fgs skyflat flat n a no fgs intflat flat n a no fgs dark dark n a no fgs id stack tracking n a no fgs id image tracking n a no fgs tracking n a no fgs tracking n a no fgs track tracking n a no fgs fineguide tracking n a no notes custom processing
| 0
|
8,745
| 11,872,370,203
|
IssuesEvent
|
2020-03-26 15:43:10
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Othermeta and source metadata should not cascade to child topicref
|
bug preprocess priority/high
|
Here is an example of the present behaviour:
The map:
``` xml
<bookmap>
<booktitle>
<mainbooktitle>Simple example</mainbooktitle>
</booktitle>
<bookmeta>
<othermeta name="this_meta" content="should not be inherited"/>
</bookmeta>
<chapter href="child.dita"/>
</bookmap>
```
The topic:
``` xml
<topic id="topic-1">
<title>The child</title>
<body>
<p>Should not have any meta.</p>
</body>
</topic>
```
The result xhtml transformation of the topic:
``` xml
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE html
PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en-us" lang="en-us">
<head><meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
<meta name="copyright" content="(C) Copyright 2015" />
<meta name="DC.rights.owner" content="(C) Copyright 2015" />
<meta name="DC.Type" content="topic" />
<meta name="DC.Title" content="The child" />
<meta name="this_meta" content="should not be inherited" />
<meta name="DC.Format" content="XHTML" />
<meta name="DC.Identifier" content="topic-1" />
<link rel="stylesheet" type="text/css" href="commonltr.css" />
<title>The child</title>
</head>
<body id="topic-1">
<h1 class="title topictitle1" id="ariaid-title1">The child</h1>
<div class="body">
<p class="p">Should not have any meta.</p>
</div>
</body>
</html>
```
As you can see, the othermeta has been inherited by the topic. But, as stated in the [dita 1.2 spec](https://dita.fluidtopics.com/book#!book;uri=urn%253Adita%253ADITA1%255C0dita-1.2-specification.ditamap%253Fv%255C2DITA1.2-spec-XHTML;breadcrumb=urn%253Adita%253ADITA1%255C0architectural_specification.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C3-base.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0ditamarkup.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0dita%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0reconciling%255C3-topic%255C3-and%255C3-map%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML), othermeta and source metadata (among others) should not cascade to children.
This bug has been introduced has part of the fix for #1086.
The fix may be to revert the incriminated commit (the one that I am unable to find).
|
1.0
|
Othermeta and source metadata should not cascade to child topicref - Here is an example of the present behaviour:
The map:
``` xml
<bookmap>
<booktitle>
<mainbooktitle>Simple example</mainbooktitle>
</booktitle>
<bookmeta>
<othermeta name="this_meta" content="should not be inherited"/>
</bookmeta>
<chapter href="child.dita"/>
</bookmap>
```
The topic:
``` xml
<topic id="topic-1">
<title>The child</title>
<body>
<p>Should not have any meta.</p>
</body>
</topic>
```
The result xhtml transformation of the topic:
``` xml
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE html
PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en-us" lang="en-us">
<head><meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
<meta name="copyright" content="(C) Copyright 2015" />
<meta name="DC.rights.owner" content="(C) Copyright 2015" />
<meta name="DC.Type" content="topic" />
<meta name="DC.Title" content="The child" />
<meta name="this_meta" content="should not be inherited" />
<meta name="DC.Format" content="XHTML" />
<meta name="DC.Identifier" content="topic-1" />
<link rel="stylesheet" type="text/css" href="commonltr.css" />
<title>The child</title>
</head>
<body id="topic-1">
<h1 class="title topictitle1" id="ariaid-title1">The child</h1>
<div class="body">
<p class="p">Should not have any meta.</p>
</div>
</body>
</html>
```
As you can see, the othermeta has been inherited by the topic. But, as stated in the [dita 1.2 spec](https://dita.fluidtopics.com/book#!book;uri=urn%253Adita%253ADITA1%255C0dita-1.2-specification.ditamap%253Fv%255C2DITA1.2-spec-XHTML;breadcrumb=urn%253Adita%253ADITA1%255C0architectural_specification.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C3-base.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0ditamarkup.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0dita%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0reconciling%255C3-topic%255C3-and%255C3-map%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML), othermeta and source metadata (among others) should not cascade to children.
This bug has been introduced has part of the fix for #1086.
The fix may be to revert the incriminated commit (the one that I am unable to find).
|
process
|
othermeta and source metadata should not cascade to child topicref here is an example of the present behaviour the map xml simple example the topic xml the child should not have any meta the result xhtml transformation of the topic xml doctype html public dtd xhtml transitional en the child the child should not have any meta as you can see the othermeta has been inherited by the topic but as stated in the othermeta and source metadata among others should not cascade to children this bug has been introduced has part of the fix for the fix may be to revert the incriminated commit the one that i am unable to find
| 1
|
1,322
| 3,874,100,994
|
IssuesEvent
|
2016-04-11 19:18:14
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
opened
|
Remove interventions named "none"
|
bug Processors
|
http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374 has a intervention named "none"
```javascript
{
"brief_summary": "",
"id": "13181266-923d-4d28-8e65-73d174314374",
"interventions": [
{
"attributes": {
"id": "94cca51c-9253-4e78-a0ed-766c21e3d083",
"name": "none"
}
}
],
"locations": [
{
"attributes": {
"id": "c2905aaf-49fa-40cb-b858-1dc2e3b484f7",
"name": "Japan",
"type": "country"
},
"role": "recruitment_countries"
}
],
"organisations": [],
"persons": [],
"problems": [
{
"attributes": {
"id": "3e09105d-a4e0-422a-b1a3-2c43c8f0ed60",
"name": "Diabetes"
}
}
],
"public_title": "Investigation of the progression of atherosclerosis and atherosclerotic risk factor in patients with diabetes",
"registration_date": "2016-04-09T00:00:00.000Z",
"url": "http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374"
}
```
|
1.0
|
Remove interventions named "none" - http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374 has a intervention named "none"
```javascript
{
"brief_summary": "",
"id": "13181266-923d-4d28-8e65-73d174314374",
"interventions": [
{
"attributes": {
"id": "94cca51c-9253-4e78-a0ed-766c21e3d083",
"name": "none"
}
}
],
"locations": [
{
"attributes": {
"id": "c2905aaf-49fa-40cb-b858-1dc2e3b484f7",
"name": "Japan",
"type": "country"
},
"role": "recruitment_countries"
}
],
"organisations": [],
"persons": [],
"problems": [
{
"attributes": {
"id": "3e09105d-a4e0-422a-b1a3-2c43c8f0ed60",
"name": "Diabetes"
}
}
],
"public_title": "Investigation of the progression of atherosclerosis and atherosclerotic risk factor in patients with diabetes",
"registration_date": "2016-04-09T00:00:00.000Z",
"url": "http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374"
}
```
|
process
|
remove interventions named none has a intervention named none javascript brief summary id interventions attributes id name none locations attributes id name japan type country role recruitment countries organisations persons problems attributes id name diabetes public title investigation of the progression of atherosclerosis and atherosclerotic risk factor in patients with diabetes registration date url
| 1
|
12,186
| 14,742,224,770
|
IssuesEvent
|
2021-01-07 11:55:26
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Keener - Recap of concerns and suggestions - Manual Invoices Billing Cycles
|
anc-process anp-1 ant-enhancement
|
In GitLab by @kdjstudios on Mar 27, 2019, 14:16
**Submitted by:** Gaylan
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/7808504
**Server:** All
**Client/Site:** All
**Account:** All
**Issue:**
I just wanted to send out a recap of our phone conversation to you and our support team.
First, we have a two concerns with how manual invoices are being handled. For manual invoices that are being created between billing cycles we would like to have the ability to change the dates to help clients better distinguish between the manual invoice and the billing cycle invoice. Also this should then help with not having the multiple invoices show on the Invoice Journal Report as the invoices could have different invoice dates. The second part of this concern is when processing a billing cycle both the manual invoice and the billing cycle invoices are being sent at the send step. We would like to have the ability to only send the Billing cycle invoice. As clients are confused when they receive two invoices on the same date. I will provide this feedback to the support team and follow up with their responses.
Second, we are wondering if there is a report that can be run for any given date and would show us how many active accounts there were on that date. I thought the Account details report would do this, but it only runs for the current date and is unable to filter to only active accounts. I will look further into this and see if there are any reports that would accomplish this.
Thank you again for the feedback and please let us know if there are any other concerns or questions you have.
|
1.0
|
Keener - Recap of concerns and suggestions - Manual Invoices Billing Cycles - In GitLab by @kdjstudios on Mar 27, 2019, 14:16
**Submitted by:** Gaylan
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/7808504
**Server:** All
**Client/Site:** All
**Account:** All
**Issue:**
I just wanted to send out a recap of our phone conversation to you and our support team.
First, we have a two concerns with how manual invoices are being handled. For manual invoices that are being created between billing cycles we would like to have the ability to change the dates to help clients better distinguish between the manual invoice and the billing cycle invoice. Also this should then help with not having the multiple invoices show on the Invoice Journal Report as the invoices could have different invoice dates. The second part of this concern is when processing a billing cycle both the manual invoice and the billing cycle invoices are being sent at the send step. We would like to have the ability to only send the Billing cycle invoice. As clients are confused when they receive two invoices on the same date. I will provide this feedback to the support team and follow up with their responses.
Second, we are wondering if there is a report that can be run for any given date and would show us how many active accounts there were on that date. I thought the Account details report would do this, but it only runs for the current date and is unable to filter to only active accounts. I will look further into this and see if there are any reports that would accomplish this.
Thank you again for the feedback and please let us know if there are any other concerns or questions you have.
|
process
|
keener recap of concerns and suggestions manual invoices billing cycles in gitlab by kdjstudios on mar submitted by gaylan helpdesk server all client site all account all issue i just wanted to send out a recap of our phone conversation to you and our support team first we have a two concerns with how manual invoices are being handled for manual invoices that are being created between billing cycles we would like to have the ability to change the dates to help clients better distinguish between the manual invoice and the billing cycle invoice also this should then help with not having the multiple invoices show on the invoice journal report as the invoices could have different invoice dates the second part of this concern is when processing a billing cycle both the manual invoice and the billing cycle invoices are being sent at the send step we would like to have the ability to only send the billing cycle invoice as clients are confused when they receive two invoices on the same date i will provide this feedback to the support team and follow up with their responses second we are wondering if there is a report that can be run for any given date and would show us how many active accounts there were on that date i thought the account details report would do this but it only runs for the current date and is unable to filter to only active accounts i will look further into this and see if there are any reports that would accomplish this thank you again for the feedback and please let us know if there are any other concerns or questions you have
| 1
|
6,226
| 9,167,379,229
|
IssuesEvent
|
2019-03-02 13:09:54
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
Cleanup Kokoro + Microsoft builds.
|
type: cleanup type: process
|
Now that `vcpkg` has the same version of `googletest` that we use we can simplify the `ci/kokoro/*.ps1` scripts. No need to use an external project.
|
1.0
|
Cleanup Kokoro + Microsoft builds. - Now that `vcpkg` has the same version of `googletest` that we use we can simplify the `ci/kokoro/*.ps1` scripts. No need to use an external project.
|
process
|
cleanup kokoro microsoft builds now that vcpkg has the same version of googletest that we use we can simplify the ci kokoro scripts no need to use an external project
| 1
|
19,628
| 25,986,178,516
|
IssuesEvent
|
2022-12-20 00:34:38
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
Analista de Negócios de TI na [Grupo LM]
|
SALVADOR REQUISITOS MODELAGEM DE PROCESSOS HELP WANTED NEGÓCIOS Stale
|
<!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Local
- Salvador
## Benefícios
- Plano de saúde
- Transporte
- Assistência Odontológica
- Ticket Refeição
## Requisitos
**Obrigatórios:**
- Formação superior completa
- Conhecimento em análise com foco no usuário, orientação ao negócio e visão de todo
- Experiência com levantamento de requisitos e modelagem de processos
- Experiência anterior na área
## Nossa empresa
- Com 45 anos de atividades, e atuação nacional, o Grupo LM (www.grupolm.com.br) é constituído pelas empresas LM Frotas, que atua no segmento de terceirização de frotas corporativas, pela concessionária Bravo Caminhões e Ônibus e também pela AuraBrasil, que opera na área de locação de máquinas e equipamentos.
## Como se candidatar
- Envie seu curriculo para marina.santos@grupolm.com.br com assunto "Analista de Negócios"
|
1.0
|
Analista de Negócios de TI na [Grupo LM] - <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Local
- Salvador
## Benefícios
- Plano de saúde
- Transporte
- Assistência Odontológica
- Ticket Refeição
## Requisitos
**Obrigatórios:**
- Formação superior completa
- Conhecimento em análise com foco no usuário, orientação ao negócio e visão de todo
- Experiência com levantamento de requisitos e modelagem de processos
- Experiência anterior na área
## Nossa empresa
- Com 45 anos de atividades, e atuação nacional, o Grupo LM (www.grupolm.com.br) é constituído pelas empresas LM Frotas, que atua no segmento de terceirização de frotas corporativas, pela concessionária Bravo Caminhões e Ônibus e também pela AuraBrasil, que opera na área de locação de máquinas e equipamentos.
## Como se candidatar
- Envie seu curriculo para marina.santos@grupolm.com.br com assunto "Analista de Negócios"
|
process
|
analista de negócios de ti na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na local salvador benefícios plano de saúde transporte assistência odontológica ticket refeição requisitos obrigatórios formação superior completa conhecimento em análise com foco no usuário orientação ao negócio e visão de todo experiência com levantamento de requisitos e modelagem de processos experiência anterior na área nossa empresa com anos de atividades e atuação nacional o grupo lm é constituído pelas empresas lm frotas que atua no segmento de terceirização de frotas corporativas pela concessionária bravo caminhões e ônibus e também pela aurabrasil que opera na área de locação de máquinas e equipamentos como se candidatar envie seu curriculo para marina santos grupolm com br com assunto analista de negócios
| 1
|
10,406
| 13,204,215,990
|
IssuesEvent
|
2020-08-14 15:30:09
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
The difference between `pipeline` and `source` is not clear.
|
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
|
The difference between `pipeline` and `source` is not clear.
In the example they seem to be the same apart from small spelling differences leaving me wonder what both of their meaning truly is.
Can they both be the same?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @ashkir
* Microsoft Alias: **ashkir**
|
1.0
|
The difference between `pipeline` and `source` is not clear. -
The difference between `pipeline` and `source` is not clear.
In the example they seem to be the same apart from small spelling differences leaving me wonder what both of their meaning truly is.
Can they both be the same?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 86285f72-9e28-da97-59bb-c29eb60f627d
* Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54
* Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @ashkir
* Microsoft Alias: **ashkir**
|
process
|
the difference between pipeline and source is not clear the difference between pipeline and source is not clear in the example they seem to be the same apart from small spelling differences leaving me wonder what both of their meaning truly is can they both be the same document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login ashkir microsoft alias ashkir
| 1
|
5,101
| 7,881,686,895
|
IssuesEvent
|
2018-06-26 19:56:29
|
ArctosDB/new-collections
|
https://api.github.com/repos/ArctosDB/new-collections
|
closed
|
Angelo State Collection Mentor(s)
|
Application in process
|
Assign collection mentor
Default = AWG Chair or Vice-Chair
AWG member can volunteer to act as primary contact, especially if they have similar collections or specific knowledge about a collection; can serve as ‘in kind support’ for collections to help offset costs
|
1.0
|
Angelo State Collection Mentor(s) - Assign collection mentor
Default = AWG Chair or Vice-Chair
AWG member can volunteer to act as primary contact, especially if they have similar collections or specific knowledge about a collection; can serve as ‘in kind support’ for collections to help offset costs
|
process
|
angelo state collection mentor s assign collection mentor default awg chair or vice chair awg member can volunteer to act as primary contact especially if they have similar collections or specific knowledge about a collection can serve as ‘in kind support’ for collections to help offset costs
| 1
|
21,488
| 29,578,222,797
|
IssuesEvent
|
2023-06-07 02:00:09
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Wed, 7 Jun 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
There is no result
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Real-Time Onboard Object Detection for Augmented Reality: Enhancing Head-Mounted Display with YOLOv8
- **Authors:** Mikołaj Łysakowski, Kamil Żywanowski, Adam Banaszczyk, Michał R. Nowicki, Piotr Skrzypczyński, Sławomir K. Tadeja
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Human-Computer Interaction (cs.HC)
- **Arxiv link:** https://arxiv.org/abs/2306.03537
- **Pdf link:** https://arxiv.org/pdf/2306.03537
- **Abstract**
This paper introduces a software architecture for real-time object detection using machine learning (ML) in an augmented reality (AR) environment. Our approach uses the recent state-of-the-art YOLOv8 network that runs onboard on the Microsoft HoloLens 2 head-mounted display (HMD). The primary motivation behind this research is to enable the application of advanced ML models for enhanced perception and situational awareness with a wearable, hands-free AR platform. We show the image processing pipeline for the YOLOv8 model and the techniques used to make it real-time on the resource-limited edge computing platform of the headset. The experimental results demonstrate that our solution achieves real-time processing without needing offloading tasks to the cloud or any other external servers while retaining satisfactory accuracy regarding the usual mAP metric and measured qualitative performance
### SDR-GAIN: A High Real-Time Occluded Pedestrian Pose Completion Method for Autonomous Driving
- **Authors:** Honghao Fu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2306.03538
- **Pdf link:** https://arxiv.org/pdf/2306.03538
- **Abstract**
To mitigate the challenges arising from partial occlusion in human pose keypoint based pedestrian detection methods , we present a novel pedestrian pose keypoint completion method called the separation and dimensionality reduction-based generative adversarial imputation networks (SDR-GAIN) . Firstly, we utilize OpenPose to estimate pedestrian poses in images. Then, we isolate the head and torso keypoints of pedestrians with incomplete keypoints due to occlusion or other factors and perform dimensionality reduction to enhance features and further unify feature distribution. Finally, we introduce two generative models based on the generative adversarial networks (GAN) framework, which incorporate Huber loss, residual structure, and L1 regularization to generate missing parts of the incomplete head and torso pose keypoints of partially occluded pedestrians, resulting in pose completion. Our experiments on MS COCO and JAAD datasets demonstrate that SDR-GAIN outperforms basic GAIN framework, interpolation methods PCHIP and MAkima, machine learning methods k-NN and MissForest in terms of pose completion task. In addition, the runtime of SDR-GAIN is approximately 0.4ms, displaying high real-time performance and significant application value in the field of autonomous driving.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
There is no result
## Keyword: RAW
### MoDAR: Using Motion Forecasting for 3D Object Detection in Point Cloud Sequences
- **Authors:** Yingwei Li, Charles R. Qi, Yin Zhou, Chenxi Liu, Dragomir Anguelov
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.03206
- **Pdf link:** https://arxiv.org/pdf/2306.03206
- **Abstract**
Occluded and long-range objects are ubiquitous and challenging for 3D object detection. Point cloud sequence data provide unique opportunities to improve such cases, as an occluded or distant object can be observed from different viewpoints or gets better visibility over time. However, the efficiency and effectiveness in encoding long-term sequence data can still be improved. In this work, we propose MoDAR, using motion forecasting outputs as a type of virtual modality, to augment LiDAR point clouds. The MoDAR modality propagates object information from temporal contexts to a target frame, represented as a set of virtual points, one for each object from a waypoint on a forecasted trajectory. A fused point cloud of both raw sensor points and the virtual points can then be fed to any off-the-shelf point-cloud based 3D object detector. Evaluated on the Waymo Open Dataset, our method significantly improves prior art detectors by using motion forecasting from extra-long sequences (e.g. 18 seconds), achieving new state of the arts, while not adding much computation overhead.
### MetaGait: Learning to Learn an Omni Sample Adaptive Representation for Gait Recognition
- **Authors:** Huanzhang Dou, Pengyi Zhang, Wei Su, Yunlong Yu, Xi Li
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2306.03445
- **Pdf link:** https://arxiv.org/pdf/2306.03445
- **Abstract**
Gait recognition, which aims at identifying individuals by their walking patterns, has recently drawn increasing research attention. However, gait recognition still suffers from the conflicts between the limited binary visual clues of the silhouette and numerous covariates with diverse scales, which brings challenges to the model's adaptiveness. In this paper, we address this conflict by developing a novel MetaGait that learns to learn an omni sample adaptive representation. Towards this goal, MetaGait injects meta-knowledge, which could guide the model to perceive sample-specific properties, into the calibration network of the attention mechanism to improve the adaptiveness from the omni-scale, omni-dimension, and omni-process perspectives. Specifically, we leverage the meta-knowledge across the entire process, where Meta Triple Attention and Meta Temporal Pooling are presented respectively to adaptively capture omni-scale dependency from spatial/channel/temporal dimensions simultaneously and to adaptively aggregate temporal information through integrating the merits of three complementary temporal aggregation methods. Extensive experiments demonstrate the state-of-the-art performance of the proposed MetaGait. On CASIA-B, we achieve rank-1 accuracy of 98.7%, 96.0%, and 89.3% under three conditions, respectively. On OU-MVLP, we achieve rank-1 accuracy of 92.4%.
### Instructive Feature Enhancement for Dichotomous Medical Image Segmentation
- **Authors:** Lian Liu, Han Zhou, Jiongquan Chen, Sijing Liu, Wenlong Shi, Dong Ni, Deng-Ping Fan, Xin Yang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.03497
- **Pdf link:** https://arxiv.org/pdf/2306.03497
- **Abstract**
Deep neural networks have been widely applied in dichotomous medical image segmentation (DMIS) of many anatomical structures in several modalities, achieving promising performance. However, existing networks tend to struggle with task-specific, heavy and complex designs to improve accuracy. They made little instructions to which feature channels would be more beneficial for segmentation, and that may be why the performance and universality of these segmentation models are hindered. In this study, we propose an instructive feature enhancement approach, namely IFE, to adaptively select feature channels with rich texture cues and strong discriminability to enhance raw features based on local curvature or global information entropy criteria. Being plug-and-play and applicable for diverse DMIS tasks, IFE encourages the model to focus on texture-rich features which are especially important for the ambiguous and challenging boundary identification, simultaneously achieving simplicity, universality, and certain interpretability. To evaluate the proposed IFE, we constructed the first large-scale DMIS dataset Cosmos55k, which contains 55,023 images from 7 modalities and 26 anatomical structures. Extensive experiments show that IFE can improve the performance of classic segmentation networks across different anatomies and modalities with only slight modifications. Code is available at https://github.com/yezi-66/IFE
### RDFC-GAN: RGB-Depth Fusion CycleGAN for Indoor Depth Completion
- **Authors:** Haowen Wang, Zhengping Che, Mingyuan Wang, Zhiyuan Xu, Xiuquan Qiao, Mengshi Qi, Feifei Feng, Jian Tang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2306.03584
- **Pdf link:** https://arxiv.org/pdf/2306.03584
- **Abstract**
The raw depth image captured by indoor depth sensors usually has an extensive range of missing depth values due to inherent limitations such as the inability to perceive transparent objects and the limited distance range. The incomplete depth map with missing values burdens many downstream vision tasks, and a rising number of depth completion methods have been proposed to alleviate this issue. While most existing methods can generate accurate dense depth maps from sparse and uniformly sampled depth maps, they are not suitable for complementing large contiguous regions of missing depth values, which is common and critical in images captured in indoor environments. To overcome these challenges, we design a novel two-branch end-to-end fusion network named RDFC-GAN, which takes a pair of RGB and incomplete depth images as input to predict a dense and completed depth map. The first branch employs an encoder-decoder structure, by adhering to the Manhattan world assumption and utilizing normal maps from RGB-D information as guidance, to regress the local dense depth values from the raw depth map. In the other branch, we propose an RGB-depth fusion CycleGAN to transfer the RGB image to the fine-grained textured depth map. We adopt adaptive fusion modules named W-AdaIN to propagate the features across the two branches, and we append a confidence fusion head to fuse the two outputs of the branches for the final depth map. Extensive experiments on NYU-Depth V2 and SUN RGB-D demonstrate that our proposed method clearly improves the depth completion performance, especially in a more realistic setting of indoor environments, with the help of our proposed pseudo depth maps in training.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Wed, 7 Jun 23 - ## Keyword: events
There is no result
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Real-Time Onboard Object Detection for Augmented Reality: Enhancing Head-Mounted Display with YOLOv8
- **Authors:** Mikołaj Łysakowski, Kamil Żywanowski, Adam Banaszczyk, Michał R. Nowicki, Piotr Skrzypczyński, Sławomir K. Tadeja
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Human-Computer Interaction (cs.HC)
- **Arxiv link:** https://arxiv.org/abs/2306.03537
- **Pdf link:** https://arxiv.org/pdf/2306.03537
- **Abstract**
This paper introduces a software architecture for real-time object detection using machine learning (ML) in an augmented reality (AR) environment. Our approach uses the recent state-of-the-art YOLOv8 network that runs onboard on the Microsoft HoloLens 2 head-mounted display (HMD). The primary motivation behind this research is to enable the application of advanced ML models for enhanced perception and situational awareness with a wearable, hands-free AR platform. We show the image processing pipeline for the YOLOv8 model and the techniques used to make it real-time on the resource-limited edge computing platform of the headset. The experimental results demonstrate that our solution achieves real-time processing without needing offloading tasks to the cloud or any other external servers while retaining satisfactory accuracy regarding the usual mAP metric and measured qualitative performance
### SDR-GAIN: A High Real-Time Occluded Pedestrian Pose Completion Method for Autonomous Driving
- **Authors:** Honghao Fu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2306.03538
- **Pdf link:** https://arxiv.org/pdf/2306.03538
- **Abstract**
To mitigate the challenges arising from partial occlusion in human pose keypoint based pedestrian detection methods , we present a novel pedestrian pose keypoint completion method called the separation and dimensionality reduction-based generative adversarial imputation networks (SDR-GAIN) . Firstly, we utilize OpenPose to estimate pedestrian poses in images. Then, we isolate the head and torso keypoints of pedestrians with incomplete keypoints due to occlusion or other factors and perform dimensionality reduction to enhance features and further unify feature distribution. Finally, we introduce two generative models based on the generative adversarial networks (GAN) framework, which incorporate Huber loss, residual structure, and L1 regularization to generate missing parts of the incomplete head and torso pose keypoints of partially occluded pedestrians, resulting in pose completion. Our experiments on MS COCO and JAAD datasets demonstrate that SDR-GAIN outperforms basic GAIN framework, interpolation methods PCHIP and MAkima, machine learning methods k-NN and MissForest in terms of pose completion task. In addition, the runtime of SDR-GAIN is approximately 0.4ms, displaying high real-time performance and significant application value in the field of autonomous driving.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
There is no result
## Keyword: RAW
### MoDAR: Using Motion Forecasting for 3D Object Detection in Point Cloud Sequences
- **Authors:** Yingwei Li, Charles R. Qi, Yin Zhou, Chenxi Liu, Dragomir Anguelov
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.03206
- **Pdf link:** https://arxiv.org/pdf/2306.03206
- **Abstract**
Occluded and long-range objects are ubiquitous and challenging for 3D object detection. Point cloud sequence data provide unique opportunities to improve such cases, as an occluded or distant object can be observed from different viewpoints or gets better visibility over time. However, the efficiency and effectiveness in encoding long-term sequence data can still be improved. In this work, we propose MoDAR, using motion forecasting outputs as a type of virtual modality, to augment LiDAR point clouds. The MoDAR modality propagates object information from temporal contexts to a target frame, represented as a set of virtual points, one for each object from a waypoint on a forecasted trajectory. A fused point cloud of both raw sensor points and the virtual points can then be fed to any off-the-shelf point-cloud based 3D object detector. Evaluated on the Waymo Open Dataset, our method significantly improves prior art detectors by using motion forecasting from extra-long sequences (e.g. 18 seconds), achieving new state of the arts, while not adding much computation overhead.
### MetaGait: Learning to Learn an Omni Sample Adaptive Representation for Gait Recognition
- **Authors:** Huanzhang Dou, Pengyi Zhang, Wei Su, Yunlong Yu, Xi Li
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2306.03445
- **Pdf link:** https://arxiv.org/pdf/2306.03445
- **Abstract**
Gait recognition, which aims at identifying individuals by their walking patterns, has recently drawn increasing research attention. However, gait recognition still suffers from the conflicts between the limited binary visual clues of the silhouette and numerous covariates with diverse scales, which brings challenges to the model's adaptiveness. In this paper, we address this conflict by developing a novel MetaGait that learns to learn an omni sample adaptive representation. Towards this goal, MetaGait injects meta-knowledge, which could guide the model to perceive sample-specific properties, into the calibration network of the attention mechanism to improve the adaptiveness from the omni-scale, omni-dimension, and omni-process perspectives. Specifically, we leverage the meta-knowledge across the entire process, where Meta Triple Attention and Meta Temporal Pooling are presented respectively to adaptively capture omni-scale dependency from spatial/channel/temporal dimensions simultaneously and to adaptively aggregate temporal information through integrating the merits of three complementary temporal aggregation methods. Extensive experiments demonstrate the state-of-the-art performance of the proposed MetaGait. On CASIA-B, we achieve rank-1 accuracy of 98.7%, 96.0%, and 89.3% under three conditions, respectively. On OU-MVLP, we achieve rank-1 accuracy of 92.4%.
### Instructive Feature Enhancement for Dichotomous Medical Image Segmentation
- **Authors:** Lian Liu, Han Zhou, Jiongquan Chen, Sijing Liu, Wenlong Shi, Dong Ni, Deng-Ping Fan, Xin Yang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.03497
- **Pdf link:** https://arxiv.org/pdf/2306.03497
- **Abstract**
Deep neural networks have been widely applied in dichotomous medical image segmentation (DMIS) of many anatomical structures in several modalities, achieving promising performance. However, existing networks tend to struggle with task-specific, heavy and complex designs to improve accuracy. They made little instructions to which feature channels would be more beneficial for segmentation, and that may be why the performance and universality of these segmentation models are hindered. In this study, we propose an instructive feature enhancement approach, namely IFE, to adaptively select feature channels with rich texture cues and strong discriminability to enhance raw features based on local curvature or global information entropy criteria. Being plug-and-play and applicable for diverse DMIS tasks, IFE encourages the model to focus on texture-rich features which are especially important for the ambiguous and challenging boundary identification, simultaneously achieving simplicity, universality, and certain interpretability. To evaluate the proposed IFE, we constructed the first large-scale DMIS dataset Cosmos55k, which contains 55,023 images from 7 modalities and 26 anatomical structures. Extensive experiments show that IFE can improve the performance of classic segmentation networks across different anatomies and modalities with only slight modifications. Code is available at https://github.com/yezi-66/IFE
### RDFC-GAN: RGB-Depth Fusion CycleGAN for Indoor Depth Completion
- **Authors:** Haowen Wang, Zhengping Che, Mingyuan Wang, Zhiyuan Xu, Xiuquan Qiao, Mengshi Qi, Feifei Feng, Jian Tang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2306.03584
- **Pdf link:** https://arxiv.org/pdf/2306.03584
- **Abstract**
The raw depth image captured by indoor depth sensors usually has an extensive range of missing depth values due to inherent limitations such as the inability to perceive transparent objects and the limited distance range. The incomplete depth map with missing values burdens many downstream vision tasks, and a rising number of depth completion methods have been proposed to alleviate this issue. While most existing methods can generate accurate dense depth maps from sparse and uniformly sampled depth maps, they are not suitable for complementing large contiguous regions of missing depth values, which is common and critical in images captured in indoor environments. To overcome these challenges, we design a novel two-branch end-to-end fusion network named RDFC-GAN, which takes a pair of RGB and incomplete depth images as input to predict a dense and completed depth map. The first branch employs an encoder-decoder structure, by adhering to the Manhattan world assumption and utilizing normal maps from RGB-D information as guidance, to regress the local dense depth values from the raw depth map. In the other branch, we propose an RGB-depth fusion CycleGAN to transfer the RGB image to the fine-grained textured depth map. We adopt adaptive fusion modules named W-AdaIN to propagate the features across the two branches, and we append a confidence fusion head to fuse the two outputs of the branches for the final depth map. Extensive experiments on NYU-Depth V2 and SUN RGB-D demonstrate that our proposed method clearly improves the depth completion performance, especially in a more realistic setting of indoor environments, with the help of our proposed pseudo depth maps in training.
## Keyword: raw image
There is no result
|
process
|
new submissions for wed jun keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp real time onboard object detection for augmented reality enhancing head mounted display with authors mikołaj łysakowski kamil żywanowski adam banaszczyk michał r nowicki piotr skrzypczyński sławomir k tadeja subjects computer vision and pattern recognition cs cv human computer interaction cs hc arxiv link pdf link abstract this paper introduces a software architecture for real time object detection using machine learning ml in an augmented reality ar environment our approach uses the recent state of the art network that runs onboard on the microsoft hololens head mounted display hmd the primary motivation behind this research is to enable the application of advanced ml models for enhanced perception and situational awareness with a wearable hands free ar platform we show the image processing pipeline for the model and the techniques used to make it real time on the resource limited edge computing platform of the headset the experimental results demonstrate that our solution achieves real time processing without needing offloading tasks to the cloud or any other external servers while retaining satisfactory accuracy regarding the usual map metric and measured qualitative performance sdr gain a high real time occluded pedestrian pose completion method for autonomous driving authors honghao fu subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract to mitigate the challenges arising from partial occlusion in human pose keypoint based pedestrian detection methods we present a novel pedestrian pose keypoint completion method called the separation and dimensionality reduction based generative adversarial imputation networks sdr gain firstly we utilize openpose to estimate pedestrian poses in images then we isolate the head and torso keypoints of pedestrians with incomplete keypoints due to occlusion or other factors and perform dimensionality reduction to enhance features and further unify feature distribution finally we introduce two generative models based on the generative adversarial networks gan framework which incorporate huber loss residual structure and regularization to generate missing parts of the incomplete head and torso pose keypoints of partially occluded pedestrians resulting in pose completion our experiments on ms coco and jaad datasets demonstrate that sdr gain outperforms basic gain framework interpolation methods pchip and makima machine learning methods k nn and missforest in terms of pose completion task in addition the runtime of sdr gain is approximately displaying high real time performance and significant application value in the field of autonomous driving keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw modar using motion forecasting for object detection in point cloud sequences authors yingwei li charles r qi yin zhou chenxi liu dragomir anguelov subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract occluded and long range objects are ubiquitous and challenging for object detection point cloud sequence data provide unique opportunities to improve such cases as an occluded or distant object can be observed from different viewpoints or gets better visibility over time however the efficiency and effectiveness in encoding long term sequence data can still be improved in this work we propose modar using motion forecasting outputs as a type of virtual modality to augment lidar point clouds the modar modality propagates object information from temporal contexts to a target frame represented as a set of virtual points one for each object from a waypoint on a forecasted trajectory a fused point cloud of both raw sensor points and the virtual points can then be fed to any off the shelf point cloud based object detector evaluated on the waymo open dataset our method significantly improves prior art detectors by using motion forecasting from extra long sequences e g seconds achieving new state of the arts while not adding much computation overhead metagait learning to learn an omni sample adaptive representation for gait recognition authors huanzhang dou pengyi zhang wei su yunlong yu xi li subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract gait recognition which aims at identifying individuals by their walking patterns has recently drawn increasing research attention however gait recognition still suffers from the conflicts between the limited binary visual clues of the silhouette and numerous covariates with diverse scales which brings challenges to the model s adaptiveness in this paper we address this conflict by developing a novel metagait that learns to learn an omni sample adaptive representation towards this goal metagait injects meta knowledge which could guide the model to perceive sample specific properties into the calibration network of the attention mechanism to improve the adaptiveness from the omni scale omni dimension and omni process perspectives specifically we leverage the meta knowledge across the entire process where meta triple attention and meta temporal pooling are presented respectively to adaptively capture omni scale dependency from spatial channel temporal dimensions simultaneously and to adaptively aggregate temporal information through integrating the merits of three complementary temporal aggregation methods extensive experiments demonstrate the state of the art performance of the proposed metagait on casia b we achieve rank accuracy of and under three conditions respectively on ou mvlp we achieve rank accuracy of instructive feature enhancement for dichotomous medical image segmentation authors lian liu han zhou jiongquan chen sijing liu wenlong shi dong ni deng ping fan xin yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep neural networks have been widely applied in dichotomous medical image segmentation dmis of many anatomical structures in several modalities achieving promising performance however existing networks tend to struggle with task specific heavy and complex designs to improve accuracy they made little instructions to which feature channels would be more beneficial for segmentation and that may be why the performance and universality of these segmentation models are hindered in this study we propose an instructive feature enhancement approach namely ife to adaptively select feature channels with rich texture cues and strong discriminability to enhance raw features based on local curvature or global information entropy criteria being plug and play and applicable for diverse dmis tasks ife encourages the model to focus on texture rich features which are especially important for the ambiguous and challenging boundary identification simultaneously achieving simplicity universality and certain interpretability to evaluate the proposed ife we constructed the first large scale dmis dataset which contains images from modalities and anatomical structures extensive experiments show that ife can improve the performance of classic segmentation networks across different anatomies and modalities with only slight modifications code is available at rdfc gan rgb depth fusion cyclegan for indoor depth completion authors haowen wang zhengping che mingyuan wang zhiyuan xu xiuquan qiao mengshi qi feifei feng jian tang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract the raw depth image captured by indoor depth sensors usually has an extensive range of missing depth values due to inherent limitations such as the inability to perceive transparent objects and the limited distance range the incomplete depth map with missing values burdens many downstream vision tasks and a rising number of depth completion methods have been proposed to alleviate this issue while most existing methods can generate accurate dense depth maps from sparse and uniformly sampled depth maps they are not suitable for complementing large contiguous regions of missing depth values which is common and critical in images captured in indoor environments to overcome these challenges we design a novel two branch end to end fusion network named rdfc gan which takes a pair of rgb and incomplete depth images as input to predict a dense and completed depth map the first branch employs an encoder decoder structure by adhering to the manhattan world assumption and utilizing normal maps from rgb d information as guidance to regress the local dense depth values from the raw depth map in the other branch we propose an rgb depth fusion cyclegan to transfer the rgb image to the fine grained textured depth map we adopt adaptive fusion modules named w adain to propagate the features across the two branches and we append a confidence fusion head to fuse the two outputs of the branches for the final depth map extensive experiments on nyu depth and sun rgb d demonstrate that our proposed method clearly improves the depth completion performance especially in a more realistic setting of indoor environments with the help of our proposed pseudo depth maps in training keyword raw image there is no result
| 1
|
817,337
| 30,637,646,630
|
IssuesEvent
|
2023-07-24 19:05:15
|
priyankarpal/ProjectsHut
|
https://api.github.com/repos/priyankarpal/ProjectsHut
|
closed
|
docs: Adding a MIT Licenses badge and some more text to Licenses section in readme
|
📄 aspect: text good first issue ✨ goal: improvement 🟨 priority: medium 🚦status: awaiting triage
|
### what's wrong with the documentation?
The readme file of the repo should have a MIT Licenses badge and some more text regarding licenses in the Licenses section to make it more professional.
### Add screenshots

### Record
- [X] I have checked the existing [issues](https://github.com/priyankarpal/ProjectsHut/issues)
- [X] I have read the [Contributing Guidelines](https://github.com/priyankarpal/ProjectsHut/blob/main/contributing.md)
- [X] I agree to follow this project's [Code of Conduct](https://github.com/priyankarpal/ProjectsHut/blob/main/CODE_OF_CONDUCT.md)
- [X] I'm a GSSoC'23 contributor
- [X] I want to work on this issue
|
1.0
|
docs: Adding a MIT Licenses badge and some more text to Licenses section in readme - ### what's wrong with the documentation?
The readme file of the repo should have a MIT Licenses badge and some more text regarding licenses in the Licenses section to make it more professional.
### Add screenshots

### Record
- [X] I have checked the existing [issues](https://github.com/priyankarpal/ProjectsHut/issues)
- [X] I have read the [Contributing Guidelines](https://github.com/priyankarpal/ProjectsHut/blob/main/contributing.md)
- [X] I agree to follow this project's [Code of Conduct](https://github.com/priyankarpal/ProjectsHut/blob/main/CODE_OF_CONDUCT.md)
- [X] I'm a GSSoC'23 contributor
- [X] I want to work on this issue
|
non_process
|
docs adding a mit licenses badge and some more text to licenses section in readme what s wrong with the documentation the readme file of the repo should have a mit licenses badge and some more text regarding licenses in the licenses section to make it more professional add screenshots record i have checked the existing i have read the i agree to follow this project s i m a gssoc contributor i want to work on this issue
| 0
|
11,243
| 14,015,299,412
|
IssuesEvent
|
2020-10-29 13:09:05
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - dc:language
|
Process - implement Term - change
|
## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): The recommended specification has been deprecated.
* Proponents (who needs this change): Everyone
Proposed new attributes of the term:
* Usage comments (recommendations regarding content, etc.):
Recommended best practice is to use a controlled vocabulary such as RFC 5646.
|
1.0
|
Change term - dc:language - ## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): The recommended specification has been deprecated.
* Proponents (who needs this change): Everyone
Proposed new attributes of the term:
* Usage comments (recommendations regarding content, etc.):
Recommended best practice is to use a controlled vocabulary such as RFC 5646.
|
process
|
change term dc language change term submitter john wieczorek justification why is this change necessary the recommended specification has been deprecated proponents who needs this change everyone proposed new attributes of the term usage comments recommendations regarding content etc recommended best practice is to use a controlled vocabulary such as rfc
| 1
|
7,939
| 8,125,087,465
|
IssuesEvent
|
2018-08-16 19:44:14
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
AMQP .NET: Broken Anchor Link
|
service-bus-messaging/svc
|
Priority: Minor
Type: Broken Link
Looks like https://github.com/MicrosoftDocs/azure-docs/commit/ed466dcd1b4f3a690d08f04288cb48f98ac37003 is what changed the header, so change line 28 from:
```
In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#unsupported-features-restrictions-and-behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP.
```
to
```
In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP.
```
Potentially reword as well since the heading no longer has "unsupported features" in it.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e28d060c-a765-ed3a-9de5-7ba9ca93d649
* Version Independent ID: e6e933ef-50d7-712e-fb4b-e7ef555fb1b6
* Content: [Azure Service Bus with .NET and AMQP 1.0](https://docs.microsoft.com/en-us/azure/service-bus-messaging/service-bus-amqp-dotnet#unsupported-features-restrictions-and-behavioral-differences)
* Content Source: [articles/service-bus-messaging/service-bus-amqp-dotnet.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-bus-messaging/service-bus-amqp-dotnet.md)
* Service: **service-bus-messaging**
* GitHub Login: @sethmanheim
* Microsoft Alias: **sethm**
|
1.0
|
AMQP .NET: Broken Anchor Link - Priority: Minor
Type: Broken Link
Looks like https://github.com/MicrosoftDocs/azure-docs/commit/ed466dcd1b4f3a690d08f04288cb48f98ac37003 is what changed the header, so change line 28 from:
```
In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#unsupported-features-restrictions-and-behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP.
```
to
```
In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP.
```
Potentially reword as well since the heading no longer has "unsupported features" in it.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e28d060c-a765-ed3a-9de5-7ba9ca93d649
* Version Independent ID: e6e933ef-50d7-712e-fb4b-e7ef555fb1b6
* Content: [Azure Service Bus with .NET and AMQP 1.0](https://docs.microsoft.com/en-us/azure/service-bus-messaging/service-bus-amqp-dotnet#unsupported-features-restrictions-and-behavioral-differences)
* Content Source: [articles/service-bus-messaging/service-bus-amqp-dotnet.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-bus-messaging/service-bus-amqp-dotnet.md)
* Service: **service-bus-messaging**
* GitHub Login: @sethmanheim
* Microsoft Alias: **sethm**
|
non_process
|
amqp net broken anchor link priority minor type broken link looks like is what changed the header so change line from in the current release there are a few api features that are not supported when using amqp these unsupported features are listed later in the section unsupported features restrictions and behavioral differences some of the advanced configuration settings also have a different meaning when using amqp to in the current release there are a few api features that are not supported when using amqp these unsupported features are listed later in the section behavioral differences some of the advanced configuration settings also have a different meaning when using amqp potentially reword as well since the heading no longer has unsupported features in it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service service bus messaging github login sethmanheim microsoft alias sethm
| 0
|
18,593
| 24,569,693,128
|
IssuesEvent
|
2022-10-13 07:39:16
|
googleapis/nodejs-certificate-manager
|
https://api.github.com/repos/googleapis/nodejs-certificate-manager
|
closed
|
Promote library to 1.0.0
|
type: bug type: process api: secretmanager
|
We should promote this library to `v1.0.0` if it is no longer pre-release.
|
1.0
|
Promote library to 1.0.0 - We should promote this library to `v1.0.0` if it is no longer pre-release.
|
process
|
promote library to we should promote this library to if it is no longer pre release
| 1
|
15,200
| 19,011,011,927
|
IssuesEvent
|
2021-11-23 09:19:12
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Move non-native Processing providers into independent plugins (Request in QGIS)
|
Processing 3.22
|
### Request for documentation
From pull request QGIS/qgis#43792
Author: @alexbruy
QGIS version: 3.22
**Move non-native Processing providers into independent plugins**
### PR Description:
## Description
Following discussion in the [QEP 226](https://github.com/qgis/QGIS-Enhancement-Proposals/issues/226) separate SAGA, GRASS and OTB providers into independent plugins. This should prevent issues with loading of the Processing plugin in cases when one of these providers is broken. Also it will make easier turning them into 3rd party plugins in the future if we will make such decision.
SAGA and GRASS plugins are enabled by default to provide the same UX as before. OTB provider can be enabled in the Plugin Manager if necessary. Enabling/disabling of these three providers now performed via the Plugin Manager like with any other plugin.
This is mostly an internal change, the only visible for users part is using Plugin Manager to activate or deactivate SAGA, GRASS and OTB providers.
### Commits tagged with [need-docs] or [FEATURE]
|
1.0
|
Move non-native Processing providers into independent plugins (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#43792
Author: @alexbruy
QGIS version: 3.22
**Move non-native Processing providers into independent plugins**
### PR Description:
## Description
Following discussion in the [QEP 226](https://github.com/qgis/QGIS-Enhancement-Proposals/issues/226) separate SAGA, GRASS and OTB providers into independent plugins. This should prevent issues with loading of the Processing plugin in cases when one of these providers is broken. Also it will make easier turning them into 3rd party plugins in the future if we will make such decision.
SAGA and GRASS plugins are enabled by default to provide the same UX as before. OTB provider can be enabled in the Plugin Manager if necessary. Enabling/disabling of these three providers now performed via the Plugin Manager like with any other plugin.
This is mostly an internal change, the only visible for users part is using Plugin Manager to activate or deactivate SAGA, GRASS and OTB providers.
### Commits tagged with [need-docs] or [FEATURE]
|
process
|
move non native processing providers into independent plugins request in qgis request for documentation from pull request qgis qgis author alexbruy qgis version move non native processing providers into independent plugins pr description description following discussion in the separate saga grass and otb providers into independent plugins this should prevent issues with loading of the processing plugin in cases when one of these providers is broken also it will make easier turning them into party plugins in the future if we will make such decision saga and grass plugins are enabled by default to provide the same ux as before otb provider can be enabled in the plugin manager if necessary enabling disabling of these three providers now performed via the plugin manager like with any other plugin this is mostly an internal change the only visible for users part is using plugin manager to activate or deactivate saga grass and otb providers commits tagged with or
| 1
|
293,025
| 8,971,918,537
|
IssuesEvent
|
2019-01-29 16:58:09
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.huffingtonpost.kr - desktop site instead of mobile site
|
browser-firefox-mobile priority-normal
|
<!-- @browser: Firefox Mobile 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver
**Browser / Version**: Firefox Mobile 66.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: font does not optimize as mobile site
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/1/e678f4cf-538b-4c66-982e-12862739276e.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190121125354</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Error: "TypeError: window.YAHOO is undefined" {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 314}]\n@https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver:314:3\n', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/ko_KR/sdk.js#xfbml=1&version=v2.10&appId=284021941754241 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-WN49HC was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=50_50 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@TopRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 370}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@x28 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 430}]', u'[JavaScript Warning: "https://imasdk.googleapis.com/js/sdkloader/ima3.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1366}]', u'[JavaScript Warning: "https://ads.admaru.com/js/ads_huffingtonpost_article_body_us.js?t=1548108177754 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1380}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ads.admaru.com/js/huffingtonpost_admaru_incontent_bigbanner.js?t=1548108177791 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1395}]', u'[console.log([Sticky Ad] > Configuration: {"stickyEnabled":true,"thresholdHeight":109,"stickyDuration":5000,"waitForAdInterval":1500,"waitForAdRetries":150,"offsetFromTop":250}) https://amp.assets.huffpost.com/assets-a380132db9db2f1d3548a6c503316cc8/js/scripts.min.js:2:2905]', u'[JavaScript Warning: "The resource at https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/en_US/fbevents.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=40_40 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.youtube.com/iframe_api was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@BottomRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1815}]', u'[JavaScript Warning: "https://d1z2jf7jlzjs58.cloudfront.net/p.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://connect.facebook.net/en_US/fbevents.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.huffingtonpost.kr - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 66.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver
**Browser / Version**: Firefox Mobile 66.0
**Operating System**: Android 8.0.0
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: font does not optimize as mobile site
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/1/e678f4cf-538b-4c66-982e-12862739276e.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190121125354</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Error: "TypeError: window.YAHOO is undefined" {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 314}]\n@https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver:314:3\n', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/ko_KR/sdk.js#xfbml=1&version=v2.10&appId=284021941754241 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-WN49HC was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=50_50 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@TopRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 370}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@x28 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 430}]', u'[JavaScript Warning: "https://imasdk.googleapis.com/js/sdkloader/ima3.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1366}]', u'[JavaScript Warning: "https://ads.admaru.com/js/ads_huffingtonpost_article_body_us.js?t=1548108177754 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1380}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ads.admaru.com/js/huffingtonpost_admaru_incontent_bigbanner.js?t=1548108177791 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1395}]', u'[console.log([Sticky Ad] > Configuration: {"stickyEnabled":true,"thresholdHeight":109,"stickyDuration":5000,"waitForAdInterval":1500,"waitForAdRetries":150,"offsetFromTop":250}) https://amp.assets.huffpost.com/assets-a380132db9db2f1d3548a6c503316cc8/js/scripts.min.js:2:2905]', u'[JavaScript Warning: "The resource at https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/en_US/fbevents.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=40_40 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.youtube.com/iframe_api was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@BottomRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1815}]', u'[JavaScript Warning: "https://d1z2jf7jlzjs58.cloudfront.net/p.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://connect.facebook.net/en_US/fbevents.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes problem type desktop site instead of mobile site description font does not optimize as mobile site steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel nightly console messages u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u n u u u u u u u u u u u u u u u u u u u u u u u u u u u u u configuration stickyenabled true thresholdheight stickyduration waitforadinterval waitforadretries offsetfromtop u u u u u u u u u u u u u u u u u u u u u u u u u u u from with ❤️
| 0
|
208,613
| 23,622,841,833
|
IssuesEvent
|
2022-08-24 22:45:03
|
kube-tarian/sigrun
|
https://api.github.com/repos/kube-tarian/sigrun
|
closed
|
CVE-2020-15257 (Medium) detected in github.com/docker/docker-v20.10.7
|
security vulnerability no-issue-activity
|
## CVE-2020-15257 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/docker/docker-v20.10.7</b></p></summary>
<p>Moby Project - a collaborative project for the container ecosystem to assemble container-based systems</p>
<p>
Dependency Hierarchy:
- github.com/google/go-containerregistry-v0.6.0 (Root Library)
- github.com/docker/cli-v20.10.7
- :x: **github.com/docker/docker-v20.10.7** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kube-tarian/sigrun/commit/df1f7d3f67826e841793324e4796be4fbd91c00f">df1f7d3f67826e841793324e4796be4fbd91c00f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
containerd is an industry-standard container runtime and is available as a daemon for Linux and Windows. In containerd before versions 1.3.9 and 1.4.3, the containerd-shim API is improperly exposed to host network containers. Access controls for the shim’s API socket verified that the connecting process had an effective UID of 0, but did not otherwise restrict access to the abstract Unix domain socket. This would allow malicious containers running in the same network namespace as the shim, with an effective UID of 0 but otherwise reduced privileges, to cause new processes to be run with elevated privileges. This vulnerability has been fixed in containerd 1.3.9 and 1.4.3. Users should update to these versions as soon as they are released. It should be noted that containers started with an old version of containerd-shim should be stopped and restarted, as running containers will continue to be vulnerable even after an upgrade. If you are not providing the ability for untrusted users to start containers in the same network namespace as the shim (typically the "host" network namespace, for example with docker run --net=host or hostNetwork: true in a Kubernetes pod) and run with an effective UID of 0, you are not vulnerable to this issue. If you are running containers with a vulnerable configuration, you can deny access to all abstract sockets with AppArmor by adding a line similar to deny unix addr=@**, to your policy. It is best practice to run containers with a reduced set of privileges, with a non-zero UID, and with isolated namespaces. The containerd maintainers strongly advise against sharing namespaces with the host. Reducing the set of isolation mechanisms used for a container necessarily increases that container's privilege, regardless of what container runtime is used for running that container.
<p>Publish Date: 2020-12-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15257>CVE-2020-15257</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4">https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4</a></p>
<p>Release Date: 2020-12-01</p>
<p>Fix Resolution: v1.3.9,v1.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-15257 (Medium) detected in github.com/docker/docker-v20.10.7 - ## CVE-2020-15257 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/docker/docker-v20.10.7</b></p></summary>
<p>Moby Project - a collaborative project for the container ecosystem to assemble container-based systems</p>
<p>
Dependency Hierarchy:
- github.com/google/go-containerregistry-v0.6.0 (Root Library)
- github.com/docker/cli-v20.10.7
- :x: **github.com/docker/docker-v20.10.7** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kube-tarian/sigrun/commit/df1f7d3f67826e841793324e4796be4fbd91c00f">df1f7d3f67826e841793324e4796be4fbd91c00f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
containerd is an industry-standard container runtime and is available as a daemon for Linux and Windows. In containerd before versions 1.3.9 and 1.4.3, the containerd-shim API is improperly exposed to host network containers. Access controls for the shim’s API socket verified that the connecting process had an effective UID of 0, but did not otherwise restrict access to the abstract Unix domain socket. This would allow malicious containers running in the same network namespace as the shim, with an effective UID of 0 but otherwise reduced privileges, to cause new processes to be run with elevated privileges. This vulnerability has been fixed in containerd 1.3.9 and 1.4.3. Users should update to these versions as soon as they are released. It should be noted that containers started with an old version of containerd-shim should be stopped and restarted, as running containers will continue to be vulnerable even after an upgrade. If you are not providing the ability for untrusted users to start containers in the same network namespace as the shim (typically the "host" network namespace, for example with docker run --net=host or hostNetwork: true in a Kubernetes pod) and run with an effective UID of 0, you are not vulnerable to this issue. If you are running containers with a vulnerable configuration, you can deny access to all abstract sockets with AppArmor by adding a line similar to deny unix addr=@**, to your policy. It is best practice to run containers with a reduced set of privileges, with a non-zero UID, and with isolated namespaces. The containerd maintainers strongly advise against sharing namespaces with the host. Reducing the set of isolation mechanisms used for a container necessarily increases that container's privilege, regardless of what container runtime is used for running that container.
<p>Publish Date: 2020-12-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15257>CVE-2020-15257</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4">https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4</a></p>
<p>Release Date: 2020-12-01</p>
<p>Fix Resolution: v1.3.9,v1.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in github com docker docker cve medium severity vulnerability vulnerable library github com docker docker moby project a collaborative project for the container ecosystem to assemble container based systems dependency hierarchy github com google go containerregistry root library github com docker cli x github com docker docker vulnerable library found in head commit a href found in base branch main vulnerability details containerd is an industry standard container runtime and is available as a daemon for linux and windows in containerd before versions and the containerd shim api is improperly exposed to host network containers access controls for the shim’s api socket verified that the connecting process had an effective uid of but did not otherwise restrict access to the abstract unix domain socket this would allow malicious containers running in the same network namespace as the shim with an effective uid of but otherwise reduced privileges to cause new processes to be run with elevated privileges this vulnerability has been fixed in containerd and users should update to these versions as soon as they are released it should be noted that containers started with an old version of containerd shim should be stopped and restarted as running containers will continue to be vulnerable even after an upgrade if you are not providing the ability for untrusted users to start containers in the same network namespace as the shim typically the host network namespace for example with docker run net host or hostnetwork true in a kubernetes pod and run with an effective uid of you are not vulnerable to this issue if you are running containers with a vulnerable configuration you can deny access to all abstract sockets with apparmor by adding a line similar to deny unix addr to your policy it is best practice to run containers with a reduced set of privileges with a non zero uid and with isolated namespaces the containerd maintainers strongly advise against sharing namespaces with the host reducing the set of isolation mechanisms used for a container necessarily increases that container s privilege regardless of what container runtime is used for running that container publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
3,255
| 6,331,988,196
|
IssuesEvent
|
2017-07-26 11:27:17
|
gaocegege/Processing.R
|
https://api.github.com/repos/gaocegege/Processing.R
|
closed
|
Support the event function
|
community/processing priority/p0 size/no-idea status/WIP type/enhancement
|
The event function is call when the event has happen.
the code is inspired by processing.py
```python
value = 0
def draw():
fill(value)
rect(25, 25, 50, 50)
def mouseClicked():
global value
if value == 0:
value = 255
else:
value = 0
```
|
1.0
|
Support the event function - The event function is call when the event has happen.
the code is inspired by processing.py
```python
value = 0
def draw():
fill(value)
rect(25, 25, 50, 50)
def mouseClicked():
global value
if value == 0:
value = 255
else:
value = 0
```
|
process
|
support the event function the event function is call when the event has happen the code is inspired by processing py python value def draw fill value rect def mouseclicked global value if value value else value
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.