Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
718,517
24,720,620,872
IssuesEvent
2022-10-20 10:22:07
YangCatalog/backend
https://api.github.com/repos/YangCatalog/backend
closed
Checkbox to disable synonyms
enhancement Priority: Low
Having the synonyms is really great. But in some cases a user may not want to use it, so having a checkbox to disable it would be great. Also when synonyms are used, it'd be good for the user to know what synonyms were used.
1.0
Checkbox to disable synonyms - Having the synonyms is really great. But in some cases a user may not want to use it, so having a checkbox to disable it would be great. Also when synonyms are used, it'd be good for the user to know what synonyms were used.
non_process
checkbox to disable synonyms having the synonyms is really great but in some cases a user may not want to use it so having a checkbox to disable it would be great also when synonyms are used it d be good for the user to know what synonyms were used
0
119,035
15,393,564,869
IssuesEvent
2021-03-03 16:51:58
grommet/hpe-design-system
https://api.github.com/repos/grommet/hpe-design-system
closed
Update Figma file to demonstrate drop suggestions on Search
Implement design
<img width="442" alt="Screen Shot 2021-02-03 at 10 44 52 AM" src="https://user-images.githubusercontent.com/62117454/106793917-ed313d80-660c-11eb-8819-28f6884ace7e.png"> Please update Figma files to align with DS site [(https://design-system.hpe.design/components/search)](url)
1.0
Update Figma file to demonstrate drop suggestions on Search - <img width="442" alt="Screen Shot 2021-02-03 at 10 44 52 AM" src="https://user-images.githubusercontent.com/62117454/106793917-ed313d80-660c-11eb-8819-28f6884ace7e.png"> Please update Figma files to align with DS site [(https://design-system.hpe.design/components/search)](url)
non_process
update figma file to demonstrate drop suggestions on search img width alt screen shot at am src please update figma files to align with ds site url
0
381,413
11,276,538,693
IssuesEvent
2020-01-14 23:33:17
googleapis/google-api-java-client-services
https://api.github.com/repos/googleapis/google-api-java-client-services
closed
Synthesis failed for acceleratedmobilepageurl
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate acceleratedmobilepageurl. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Checking out files: 24% (15878/65361) Checking out files: 25% (16341/65361) Checking out files: 26% (16994/65361) Checking out files: 27% (17648/65361) Checking out files: 28% (18302/65361) Checking out files: 29% (18955/65361) Checking out files: 30% (19609/65361) Checking out files: 31% (20262/65361) Checking out files: 32% (20916/65361) Checking out files: 33% (21570/65361) Checking out files: 34% (22223/65361) Checking out files: 35% (22877/65361) Checking out files: 36% (23530/65361) Checking out files: 37% (24184/65361) Checking out files: 38% (24838/65361) Checking out files: 39% (25491/65361) Checking out files: 40% (26145/65361) Checking out files: 41% (26799/65361) Checking out files: 42% (27452/65361) Checking out files: 43% (28106/65361) Checking out files: 44% (28759/65361) Checking out files: 45% (29413/65361) Checking out files: 46% (30067/65361) Checking out files: 47% (30720/65361) Checking out files: 48% (31374/65361) Checking out files: 49% (32027/65361) Checking out files: 50% (32681/65361) Checking out files: 51% (33335/65361) Checking out files: 52% (33988/65361) Checking out files: 53% (34642/65361) Checking out files: 53% (35269/65361) Checking out files: 54% (35295/65361) Checking out files: 55% (35949/65361) Checking out files: 56% (36603/65361) Checking out files: 57% (37256/65361) Checking out files: 58% (37910/65361) Checking out files: 59% (38563/65361) Checking out files: 60% (39217/65361) Checking out files: 61% (39871/65361) Checking out files: 62% (40524/65361) Checking out files: 63% (41178/65361) Checking out files: 64% (41832/65361) Checking out files: 65% (42485/65361) Checking out files: 66% (43139/65361) Checking out files: 67% (43792/65361) Checking out files: 68% (44446/65361) Checking out files: 69% (45100/65361) Checking out files: 70% (45753/65361) Checking out files: 71% (46407/65361) Checking out files: 72% (47060/65361) Checking out files: 73% (47714/65361) Checking out files: 74% (48368/65361) Checking out files: 75% (49021/65361) Checking out files: 76% (49675/65361) Checking out files: 77% (50328/65361) Checking out files: 78% (50982/65361) Checking out files: 79% (51636/65361) Checking out files: 80% (52289/65361) Checking out files: 81% (52943/65361) Checking out files: 82% (53597/65361) Checking out files: 82% (53676/65361) Checking out files: 83% (54250/65361) Checking out files: 84% (54904/65361) Checking out files: 85% (55557/65361) Checking out files: 86% (56211/65361) Checking out files: 87% (56865/65361) Checking out files: 88% (57518/65361) Checking out files: 89% (58172/65361) Checking out files: 90% (58825/65361) Checking out files: 91% (59479/65361) Checking out files: 92% (60133/65361) Checking out files: 93% (60786/65361) Checking out files: 94% (61440/65361) Checking out files: 95% (62093/65361) Checking out files: 96% (62747/65361) Checking out files: 97% (63401/65361) Checking out files: 98% (64054/65361) Checking out files: 99% (64708/65361) Checking out files: 100% (65361/65361) Checking out files: 100% (65361/65361), done. Switched to branch 'autosynth-acceleratedmobilepageurl' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
1.0
Synthesis failed for acceleratedmobilepageurl - Hello! Autosynth couldn't regenerate acceleratedmobilepageurl. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Checking out files: 24% (15878/65361) Checking out files: 25% (16341/65361) Checking out files: 26% (16994/65361) Checking out files: 27% (17648/65361) Checking out files: 28% (18302/65361) Checking out files: 29% (18955/65361) Checking out files: 30% (19609/65361) Checking out files: 31% (20262/65361) Checking out files: 32% (20916/65361) Checking out files: 33% (21570/65361) Checking out files: 34% (22223/65361) Checking out files: 35% (22877/65361) Checking out files: 36% (23530/65361) Checking out files: 37% (24184/65361) Checking out files: 38% (24838/65361) Checking out files: 39% (25491/65361) Checking out files: 40% (26145/65361) Checking out files: 41% (26799/65361) Checking out files: 42% (27452/65361) Checking out files: 43% (28106/65361) Checking out files: 44% (28759/65361) Checking out files: 45% (29413/65361) Checking out files: 46% (30067/65361) Checking out files: 47% (30720/65361) Checking out files: 48% (31374/65361) Checking out files: 49% (32027/65361) Checking out files: 50% (32681/65361) Checking out files: 51% (33335/65361) Checking out files: 52% (33988/65361) Checking out files: 53% (34642/65361) Checking out files: 53% (35269/65361) Checking out files: 54% (35295/65361) Checking out files: 55% (35949/65361) Checking out files: 56% (36603/65361) Checking out files: 57% (37256/65361) Checking out files: 58% (37910/65361) Checking out files: 59% (38563/65361) Checking out files: 60% (39217/65361) Checking out files: 61% (39871/65361) Checking out files: 62% (40524/65361) Checking out files: 63% (41178/65361) Checking out files: 64% (41832/65361) Checking out files: 65% (42485/65361) Checking out files: 66% (43139/65361) Checking out files: 67% (43792/65361) Checking out files: 68% (44446/65361) Checking out files: 69% (45100/65361) Checking out files: 70% (45753/65361) Checking out files: 71% (46407/65361) Checking out files: 72% (47060/65361) Checking out files: 73% (47714/65361) Checking out files: 74% (48368/65361) Checking out files: 75% (49021/65361) Checking out files: 76% (49675/65361) Checking out files: 77% (50328/65361) Checking out files: 78% (50982/65361) Checking out files: 79% (51636/65361) Checking out files: 80% (52289/65361) Checking out files: 81% (52943/65361) Checking out files: 82% (53597/65361) Checking out files: 82% (53676/65361) Checking out files: 83% (54250/65361) Checking out files: 84% (54904/65361) Checking out files: 85% (55557/65361) Checking out files: 86% (56211/65361) Checking out files: 87% (56865/65361) Checking out files: 88% (57518/65361) Checking out files: 89% (58172/65361) Checking out files: 90% (58825/65361) Checking out files: 91% (59479/65361) Checking out files: 92% (60133/65361) Checking out files: 93% (60786/65361) Checking out files: 94% (61440/65361) Checking out files: 95% (62093/65361) Checking out files: 96% (62747/65361) Checking out files: 97% (63401/65361) Checking out files: 98% (64054/65361) Checking out files: 99% (64708/65361) Checking out files: 100% (65361/65361) Checking out files: 100% (65361/65361), done. Switched to branch 'autosynth-acceleratedmobilepageurl' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/40f694d4-43de-41f0-b993-f4694e4a45de).
non_process
synthesis failed for acceleratedmobilepageurl hello autosynth couldn t regenerate acceleratedmobilepageurl broken heart here s the output from running synth py cloning into working repo checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files checking out files done switched to branch autosynth acceleratedmobilepageurl traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log
0
14,041
16,849,514,532
IssuesEvent
2021-06-20 07:54:41
log2timeline/plaso
https://api.github.com/repos/log2timeline/plaso
opened
Add support for preprocessing plugins to indicate they dependent on another plugin
enhancement preprocessing
To resolve Windows localized time zone names WindowsAvailableTimeZones needs to run before WindowsTimezone (https://github.com/log2timeline/plaso/issues/2673). Change preprocessing plugins to define such dependencies.
1.0
Add support for preprocessing plugins to indicate they dependent on another plugin - To resolve Windows localized time zone names WindowsAvailableTimeZones needs to run before WindowsTimezone (https://github.com/log2timeline/plaso/issues/2673). Change preprocessing plugins to define such dependencies.
process
add support for preprocessing plugins to indicate they dependent on another plugin to resolve windows localized time zone names windowsavailabletimezones needs to run before windowstimezone change preprocessing plugins to define such dependencies
1
14,804
18,104,027,647
IssuesEvent
2021-09-22 17:06:15
GoogleCloudPlatform/php-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/php-docs-samples
closed
storage: separate each sample into a separate file
priority: p2 type: process help wanted api: storage samples
Each sample should be in a standalone file. This workitem is to track that work.
1.0
storage: separate each sample into a separate file - Each sample should be in a standalone file. This workitem is to track that work.
process
storage separate each sample into a separate file each sample should be in a standalone file this workitem is to track that work
1
12,815
15,190,014,434
IssuesEvent
2021-02-15 17:14:48
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
Process does not run successfully on HP UX
Bug Process Status: Needs Review
**Symfony version(s) affected**: Symfony\Process 2.8 **Description** NOTE: This is running on HP UX B11.31 which may not be worth your time as it's about to be discontinued. Still, this is my poison. I'm using Gitlist which depends on gitter which depends on Symfony\Process. Process is used to run a git log command, and while I can run this command in the commandline without errors, it does not actually run using Symfony\Process. Or at least `isSuccessful()` returns `false`. Other git commands run fine. **How to reproduce** ```php $process = new Process("'/usr/local/bin/git' -c \"color.ui\"=false log --skip=0 --max-count=15 --pretty=format:\"%H%h%T%P%aN%aE%at%cN%cE%ct\" master", "/sw/git/des_mnu"); $process->setTimeout(180); $process->run(); if (!$process->isSuccessful()) { die("Rattle"); } ``` This always fails. Other commands, like ``` '/usr/local/bin/git' -c \"color.ui\"=false branch ``` run fine. Note that running this command just gives multiple plain-text lines with commit info. Not sure why it is failing. **Possible Solution** No real idea. **Additional context** OS: An antiquated HP UX B11.31
1.0
Process does not run successfully on HP UX - **Symfony version(s) affected**: Symfony\Process 2.8 **Description** NOTE: This is running on HP UX B11.31 which may not be worth your time as it's about to be discontinued. Still, this is my poison. I'm using Gitlist which depends on gitter which depends on Symfony\Process. Process is used to run a git log command, and while I can run this command in the commandline without errors, it does not actually run using Symfony\Process. Or at least `isSuccessful()` returns `false`. Other git commands run fine. **How to reproduce** ```php $process = new Process("'/usr/local/bin/git' -c \"color.ui\"=false log --skip=0 --max-count=15 --pretty=format:\"%H%h%T%P%aN%aE%at%cN%cE%ct\" master", "/sw/git/des_mnu"); $process->setTimeout(180); $process->run(); if (!$process->isSuccessful()) { die("Rattle"); } ``` This always fails. Other commands, like ``` '/usr/local/bin/git' -c \"color.ui\"=false branch ``` run fine. Note that running this command just gives multiple plain-text lines with commit info. Not sure why it is failing. **Possible Solution** No real idea. **Additional context** OS: An antiquated HP UX B11.31
process
process does not run successfully on hp ux symfony version s affected symfony process description note this is running on hp ux which may not be worth your time as it s about to be discontinued still this is my poison i m using gitlist which depends on gitter which depends on symfony process process is used to run a git log command and while i can run this command in the commandline without errors it does not actually run using symfony process or at least issuccessful returns false other git commands run fine how to reproduce php process new process usr local bin git c color ui false log skip max count pretty format h h t p an ae at cn ce ct master sw git des mnu process settimeout process run if process issuccessful die rattle this always fails other commands like usr local bin git c color ui false branch run fine note that running this command just gives multiple plain text lines with commit info not sure why it is failing possible solution no real idea additional context os an antiquated hp ux
1
35,585
17,140,578,731
IssuesEvent
2021-07-13 09:09:46
Yoast/wordpress-seo
https://api.github.com/repos/Yoast/wordpress-seo
closed
fill_cache creates high server load
Yoast: Management component: indexables component: performance severity: minor
<!-- Please use this template when creating an issue. - Please check the boxes after you've created your issue. - Please use the latest version of Yoast SEO.--> * [x] I've read and understood the [contribution guidelines](https://github.com/Yoast/wordpress-seo/blob/trunk/.github/CONTRIBUTING.md). * [x] I've searched for any related issues and avoided creating a duplicate issue. ### Please give us a description of what happened. After upgrade from Yoast SEO from 13.5 to 16.0.2 experiencing high server load. ### Please describe what you expected to happen and why. The reason is very slow SQL query being executed on every ajax request: ``` SELECT SQL_CALC_FOUND_ROWS wp_posts.ID FROM wp_posts WHERE 1=1 AND wp_posts.post_type = 'post' AND (wp_posts.post_status = 'publish' OR wp_posts.post_status = 'future' OR wp_posts.post_status = 'draft' OR wp_posts.post_status = 'pending') ORDER BY wp_posts.post_date DESC LIMIT 0, 10 ``` This query takes 0.5 sec on our server and EXPLAIN shows: ``` +------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+ | id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | +------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+ | 1 | SIMPLE | wp_posts | range | type_status_date | type_status_date | 164 | NULL | 453574 | Using where; Using index; Using filesort | +------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+ ``` It is being called from: ``` do_action('admin_init'), WP_Hook->do_action, WP_Hook->apply_filters, Yoast\WP\SEO\Integrations\Admin\Admin_Columns_Cache_Integration->fill_cache, WP_Query->get_posts ``` Temporary workaround is to comment out fill_cache call in function register_hooks() in wp-content/plugins/wordpress-seo/src/integrations/admin/admin-columns-cache-integration.php: ``` // \add_action( 'admin_init', [ $this, 'fill_cache' ] ); ``` but actually this line should be checked: ``` $posts = empty( $wp_query->posts ) ? $wp_query->get_posts() : $wp_query->posts; ``` #### Used versions * WordPress version: 5.7 * Yoast SEO version: 16.0.2
True
fill_cache creates high server load - <!-- Please use this template when creating an issue. - Please check the boxes after you've created your issue. - Please use the latest version of Yoast SEO.--> * [x] I've read and understood the [contribution guidelines](https://github.com/Yoast/wordpress-seo/blob/trunk/.github/CONTRIBUTING.md). * [x] I've searched for any related issues and avoided creating a duplicate issue. ### Please give us a description of what happened. After upgrade from Yoast SEO from 13.5 to 16.0.2 experiencing high server load. ### Please describe what you expected to happen and why. The reason is very slow SQL query being executed on every ajax request: ``` SELECT SQL_CALC_FOUND_ROWS wp_posts.ID FROM wp_posts WHERE 1=1 AND wp_posts.post_type = 'post' AND (wp_posts.post_status = 'publish' OR wp_posts.post_status = 'future' OR wp_posts.post_status = 'draft' OR wp_posts.post_status = 'pending') ORDER BY wp_posts.post_date DESC LIMIT 0, 10 ``` This query takes 0.5 sec on our server and EXPLAIN shows: ``` +------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+ | id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | +------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+ | 1 | SIMPLE | wp_posts | range | type_status_date | type_status_date | 164 | NULL | 453574 | Using where; Using index; Using filesort | +------+-------------+----------+-------+------------------+------------------+---------+------+--------+------------------------------------------+ ``` It is being called from: ``` do_action('admin_init'), WP_Hook->do_action, WP_Hook->apply_filters, Yoast\WP\SEO\Integrations\Admin\Admin_Columns_Cache_Integration->fill_cache, WP_Query->get_posts ``` Temporary workaround is to comment out fill_cache call in function register_hooks() in wp-content/plugins/wordpress-seo/src/integrations/admin/admin-columns-cache-integration.php: ``` // \add_action( 'admin_init', [ $this, 'fill_cache' ] ); ``` but actually this line should be checked: ``` $posts = empty( $wp_query->posts ) ? $wp_query->get_posts() : $wp_query->posts; ``` #### Used versions * WordPress version: 5.7 * Yoast SEO version: 16.0.2
non_process
fill cache creates high server load please use this template when creating an issue please check the boxes after you ve created your issue please use the latest version of yoast seo i ve read and understood the i ve searched for any related issues and avoided creating a duplicate issue please give us a description of what happened after upgrade from yoast seo from to experiencing high server load please describe what you expected to happen and why the reason is very slow sql query being executed on every ajax request select sql calc found rows wp posts id from wp posts where and wp posts post type post and wp posts post status publish or wp posts post status future or wp posts post status draft or wp posts post status pending order by wp posts post date desc limit this query takes sec on our server and explain shows id select type table type possible keys key key len ref rows extra simple wp posts range type status date type status date null using where using index using filesort it is being called from do action admin init wp hook do action wp hook apply filters yoast wp seo integrations admin admin columns cache integration fill cache wp query get posts temporary workaround is to comment out fill cache call in function register hooks in wp content plugins wordpress seo src integrations admin admin columns cache integration php add action admin init but actually this line should be checked posts empty wp query posts wp query get posts wp query posts used versions wordpress version yoast seo version
0
9,424
12,417,445,228
IssuesEvent
2020-05-22 20:44:05
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Which tasks work out-of-the-box with environments?
Pri1 devops-cicd-process/tech devops/prod
So far the only task I've been able to find is "IIS web app deploy" (IISWebAppDeploymentOnMachineGroup). I need a task for doing file copy to the target server(s) of the environment, but the ones I've looked at all require specifying machine host name/IP address as well as admin credentials. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Which tasks work out-of-the-box with environments? - So far the only task I've been able to find is "IIS web app deploy" (IISWebAppDeploymentOnMachineGroup). I need a task for doing file copy to the target server(s) of the environment, but the ones I've looked at all require specifying machine host name/IP address as well as admin credentials. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
which tasks work out of the box with environments so far the only task i ve been able to find is iis web app deploy iiswebappdeploymentonmachinegroup i need a task for doing file copy to the target server s of the environment but the ones i ve looked at all require specifying machine host name ip address as well as admin credentials document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
2,980
5,965,663,282
IssuesEvent
2017-05-30 12:14:35
openvstorage/alba
https://api.github.com/repos/openvstorage/alba
closed
Failure "chunk 0 failed" must be more detailed when environment is full
process_wontfix type_enhancement
Failure "chunk 0 failed" must be more detailed when environment is full Alba version: 1.3.6 Situation: the environment got full and the policy (12/4/14/2 - 10 nodes) could not be satisfied (disks 99% on 5 nodes). I checked the logfiles and found the below error in the alba-proxy. ``` (Failure "chunk 0 failed") ``` Detailed logview: ``` 2017-02-20 10:32:43 981517 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654778 - error - Unexpected exception in proxy while handling request: (Failure "chunk 0 failed"); backtrace:; Raised at file "queue.ml", line 68, characters 17-22; Called from file "src/tools/lwt_pool2.ml", line 98, characters 25-46 2017-02-20 10:32:43 981637 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654779 - error - Request ApplySequence ("75450521-1cde-4f82-88bf-2332ffff016d",false,[(Nsm_model.Assert.ObjectHasChecksum ("owner_tag",; Sha1 c66c65175fecc3103b3b587be9b5b230889c8628)); ],[(Proxy_protocol.Protocol.Update.UploadObjectFromFile; ("00_00000049_00",; "/mnt/ssd4/vmstor_write_sco_1/75450521-1cde-4f82-88bf-2332ffff016d/00_00000049_00",; (Some Crc32c 0xdb988818))); ]) errored and took 0.189443 ``` The first "chunk 0 failed" error started right after lot of the following messages: ``` 2017-02-19 20:04:18 686779 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1408944 - warning - fragment upload failed:Asd_protocol.Protocol.Error.Exn(1) ``` ASD on one of the nodes that got filled up to 99%: ``` 2017-02-19 20:04:22 994301 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35688 - info - returning error Asd_protocol.Protocol.Error.Full 2017-02-19 20:04:23 302203 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35690 - info - returning error Asd_protocol.Protocol.Error.Full 2017-02-19 20:04:23 848663 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35691 - info - returning error Asd_protocol.Protocol.Error.Full ```
1.0
Failure "chunk 0 failed" must be more detailed when environment is full - Failure "chunk 0 failed" must be more detailed when environment is full Alba version: 1.3.6 Situation: the environment got full and the policy (12/4/14/2 - 10 nodes) could not be satisfied (disks 99% on 5 nodes). I checked the logfiles and found the below error in the alba-proxy. ``` (Failure "chunk 0 failed") ``` Detailed logview: ``` 2017-02-20 10:32:43 981517 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654778 - error - Unexpected exception in proxy while handling request: (Failure "chunk 0 failed"); backtrace:; Raised at file "queue.ml", line 68, characters 17-22; Called from file "src/tools/lwt_pool2.ml", line 98, characters 25-46 2017-02-20 10:32:43 981637 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1654779 - error - Request ApplySequence ("75450521-1cde-4f82-88bf-2332ffff016d",false,[(Nsm_model.Assert.ObjectHasChecksum ("owner_tag",; Sha1 c66c65175fecc3103b3b587be9b5b230889c8628)); ],[(Proxy_protocol.Protocol.Update.UploadObjectFromFile; ("00_00000049_00",; "/mnt/ssd4/vmstor_write_sco_1/75450521-1cde-4f82-88bf-2332ffff016d/00_00000049_00",; (Some Crc32c 0xdb988818))); ]) errored and took 0.189443 ``` The first "chunk 0 failed" error started right after lot of the following messages: ``` 2017-02-19 20:04:18 686779 +0100 - stor-01.be-g8-3 - 5973/0 - alba/proxy - 1408944 - warning - fragment upload failed:Asd_protocol.Protocol.Error.Exn(1) ``` ASD on one of the nodes that got filled up to 99%: ``` 2017-02-19 20:04:22 994301 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35688 - info - returning error Asd_protocol.Protocol.Error.Full 2017-02-19 20:04:23 302203 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35690 - info - returning error Asd_protocol.Protocol.Error.Full 2017-02-19 20:04:23 848663 +0100 - cpu-01.be-g8-3 - 11150/0 - alba/asd - 35691 - info - returning error Asd_protocol.Protocol.Error.Full ```
process
failure chunk failed must be more detailed when environment is full failure chunk failed must be more detailed when environment is full alba version situation the environment got full and the policy nodes could not be satisfied disks on nodes i checked the logfiles and found the below error in the alba proxy failure chunk failed detailed logview stor be alba proxy error unexpected exception in proxy while handling request failure chunk failed backtrace raised at file queue ml line characters called from file src tools lwt ml line characters stor be alba proxy error request applysequence false errored and took the first chunk failed error started right after lot of the following messages stor be alba proxy warning fragment upload failed asd protocol protocol error exn asd on one of the nodes that got filled up to cpu be alba asd info returning error asd protocol protocol error full cpu be alba asd info returning error asd protocol protocol error full cpu be alba asd info returning error asd protocol protocol error full
1
7,967
11,147,910,893
IssuesEvent
2019-12-23 14:03:51
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
closed
`prisma2 init` shortly shows `yarn install` stuff
bug/2-confirmed kind/bug process/candidate topic: cli-init
![yarn_install](https://user-images.githubusercontent.com/183673/65871192-2b19b400-e37e-11e9-8555-1322b2628707.gif) (Very possible this is a Windows only thing, ran this in `cmd.exe` - I have never seen this before preview-13 though) `prisma2@2.0.0-preview-13.1, binary version: 14576370b55d5152d9cb90f4f509bfa56760dbbd`
1.0
`prisma2 init` shortly shows `yarn install` stuff - ![yarn_install](https://user-images.githubusercontent.com/183673/65871192-2b19b400-e37e-11e9-8555-1322b2628707.gif) (Very possible this is a Windows only thing, ran this in `cmd.exe` - I have never seen this before preview-13 though) `prisma2@2.0.0-preview-13.1, binary version: 14576370b55d5152d9cb90f4f509bfa56760dbbd`
process
init shortly shows yarn install stuff very possible this is a windows only thing ran this in cmd exe i have never seen this before preview though preview binary version
1
136,613
19,899,403,693
IssuesEvent
2022-01-25 05:28:25
chapel-lang/chapel
https://api.github.com/repos/chapel-lang/chapel
closed
Implement 'atomic' statements and transactional memory
type: Design area: Language
In Chapel's language specification, in particular section 25.10, it makes mention of `atomic` statements. It does not specify that `atomic` blocks need to use transactional memory, but it would benefit greatly from doing so. I've been experimenting with [STM](https://github.com/LouisJenkinsCS/Chapel-Atomic-Objects/blob/master/STM.chpl) in Chapel in the past, but unfortunately due to the lack of pointer-arithmetic on wide pointers (I.E pointers to remote memory) and the lack of support for atomic operations on class instances (which I have also experimented with in the past in issue #6663 and other experiments I won't mention), it isn't really possible to go further with the idea from the module level. While [others have attempted it from the lower levels](https://github.com/chapel-lang/chapel-attic/tree/collaborations/gtm), Chapel still lacks any transactional memory support. I believe that this is something that should be done within the next few releases... I do plan on digging into it myself within the next few months after I am no longer under contract, but it would be nice to know if there is any potential discussion to be had.
1.0
Implement 'atomic' statements and transactional memory - In Chapel's language specification, in particular section 25.10, it makes mention of `atomic` statements. It does not specify that `atomic` blocks need to use transactional memory, but it would benefit greatly from doing so. I've been experimenting with [STM](https://github.com/LouisJenkinsCS/Chapel-Atomic-Objects/blob/master/STM.chpl) in Chapel in the past, but unfortunately due to the lack of pointer-arithmetic on wide pointers (I.E pointers to remote memory) and the lack of support for atomic operations on class instances (which I have also experimented with in the past in issue #6663 and other experiments I won't mention), it isn't really possible to go further with the idea from the module level. While [others have attempted it from the lower levels](https://github.com/chapel-lang/chapel-attic/tree/collaborations/gtm), Chapel still lacks any transactional memory support. I believe that this is something that should be done within the next few releases... I do plan on digging into it myself within the next few months after I am no longer under contract, but it would be nice to know if there is any potential discussion to be had.
non_process
implement atomic statements and transactional memory in chapel s language specification in particular section it makes mention of atomic statements it does not specify that atomic blocks need to use transactional memory but it would benefit greatly from doing so i ve been experimenting with in chapel in the past but unfortunately due to the lack of pointer arithmetic on wide pointers i e pointers to remote memory and the lack of support for atomic operations on class instances which i have also experimented with in the past in issue and other experiments i won t mention it isn t really possible to go further with the idea from the module level while chapel still lacks any transactional memory support i believe that this is something that should be done within the next few releases i do plan on digging into it myself within the next few months after i am no longer under contract but it would be nice to know if there is any potential discussion to be had
0
317,231
27,221,149,177
IssuesEvent
2023-02-21 05:30:29
wpfoodmanager/wp-food-manager
https://api.github.com/repos/wpfoodmanager/wp-food-manager
closed
Food category and Type listing design is not proper
Bug In Testing
Food category and type listing page design is not proper. Make it proper like food listing ![image](https://user-images.githubusercontent.com/121149500/217161536-1749a40e-6877-4297-ac9f-468d9663a60b.png) ![image](https://user-images.githubusercontent.com/121149500/217161567-cd7cb596-8d13-493e-93fc-469795b72e17.png) ![image](https://user-images.githubusercontent.com/121149500/217161645-d7d03f85-28e9-458b-a883-dd12714c830e.png) ![image](https://user-images.githubusercontent.com/121149500/217161705-7e6c38a2-6b56-43e5-ad68-bd0108e2f5f8.png)
1.0
Food category and Type listing design is not proper - Food category and type listing page design is not proper. Make it proper like food listing ![image](https://user-images.githubusercontent.com/121149500/217161536-1749a40e-6877-4297-ac9f-468d9663a60b.png) ![image](https://user-images.githubusercontent.com/121149500/217161567-cd7cb596-8d13-493e-93fc-469795b72e17.png) ![image](https://user-images.githubusercontent.com/121149500/217161645-d7d03f85-28e9-458b-a883-dd12714c830e.png) ![image](https://user-images.githubusercontent.com/121149500/217161705-7e6c38a2-6b56-43e5-ad68-bd0108e2f5f8.png)
non_process
food category and type listing design is not proper food category and type listing page design is not proper make it proper like food listing
0
12,770
15,148,878,639
IssuesEvent
2021-02-11 11:15:07
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
White points when demosaicing with 'local average'
bug: pending reproduce: peculiar scope: image processing
**Describe the bug** I was working on an image from PlayRaw at pixls.us. After switching demosaicing from PPG to AMaZE many big white spots appeared all over the image. Happens when 'match greens' is set to 'local average' or 'full and local average' with demosaicing methods 'AMaZe', 'PPG' and 'VNG4' ![amaze](https://user-images.githubusercontent.com/49001045/58758882-7af88380-8522-11e9-8676-d36cff5fa05d.jpg) XMP: (https://github.com/darktable-org/darktable/files/3244592/20190531-092259-LackawannaStatePark_BullHillTrail-0037.cr2.zip) RAW file: https://discuss.pixls.us/t/play-raw-salamander-in-the-woods/ In exported images the white points are small but still present **To Reproduce** * Load RAW file with attached XMP * Set 'match greens' in demosaicing to 'local average' or 'full and local average' **Expected behavior** Since the Darktable manual does not tell anything about white points that appear, these white spots should not be there **Platform (please complete the following information):** Darktable 2.6.2 from PPA on Ubuntu 18.04
1.0
White points when demosaicing with 'local average' - **Describe the bug** I was working on an image from PlayRaw at pixls.us. After switching demosaicing from PPG to AMaZE many big white spots appeared all over the image. Happens when 'match greens' is set to 'local average' or 'full and local average' with demosaicing methods 'AMaZe', 'PPG' and 'VNG4' ![amaze](https://user-images.githubusercontent.com/49001045/58758882-7af88380-8522-11e9-8676-d36cff5fa05d.jpg) XMP: (https://github.com/darktable-org/darktable/files/3244592/20190531-092259-LackawannaStatePark_BullHillTrail-0037.cr2.zip) RAW file: https://discuss.pixls.us/t/play-raw-salamander-in-the-woods/ In exported images the white points are small but still present **To Reproduce** * Load RAW file with attached XMP * Set 'match greens' in demosaicing to 'local average' or 'full and local average' **Expected behavior** Since the Darktable manual does not tell anything about white points that appear, these white spots should not be there **Platform (please complete the following information):** Darktable 2.6.2 from PPA on Ubuntu 18.04
process
white points when demosaicing with local average describe the bug i was working on an image from playraw at pixls us after switching demosaicing from ppg to amaze many big white spots appeared all over the image happens when match greens is set to local average or full and local average with demosaicing methods amaze ppg and xmp raw file in exported images the white points are small but still present to reproduce load raw file with attached xmp set match greens in demosaicing to local average or full and local average expected behavior since the darktable manual does not tell anything about white points that appear these white spots should not be there platform please complete the following information darktable from ppa on ubuntu
1
20,569
27,229,965,131
IssuesEvent
2023-02-21 12:31:03
corona-warn-app/cwa-wishlist
https://api.github.com/repos/corona-warn-app/cwa-wishlist
closed
Clear statement after exposure submission that others were warned
enhancement mirrored-to-jira Test/Share process
## Current Implementation Currently, the app only makes vague statements that others were warned after the "Warn others" process was finished successfully. Actually, the only thing I found was this message: "Thanks to your support, other people can now be warned and respond appropriately.". ## Suggested Enhancement Make a clear statement whether others were already warned or not, like. "The process to warn others was finished successfully and other people are now warned.". ## Expected Benefits I wrote with quite a few users on Twitter via DM who were unsure if the app did warn others already, or if they still have to do anything. This is especially a problem if the app warned automatically in the background. --- Internal Tracking ID: [EXPOSUREAPP-13576](https://jira-ibs.wbs.net.sap/browse/EXPOSUREAPP-13576)
1.0
Clear statement after exposure submission that others were warned - ## Current Implementation Currently, the app only makes vague statements that others were warned after the "Warn others" process was finished successfully. Actually, the only thing I found was this message: "Thanks to your support, other people can now be warned and respond appropriately.". ## Suggested Enhancement Make a clear statement whether others were already warned or not, like. "The process to warn others was finished successfully and other people are now warned.". ## Expected Benefits I wrote with quite a few users on Twitter via DM who were unsure if the app did warn others already, or if they still have to do anything. This is especially a problem if the app warned automatically in the background. --- Internal Tracking ID: [EXPOSUREAPP-13576](https://jira-ibs.wbs.net.sap/browse/EXPOSUREAPP-13576)
process
clear statement after exposure submission that others were warned current implementation currently the app only makes vague statements that others were warned after the warn others process was finished successfully actually the only thing i found was this message thanks to your support other people can now be warned and respond appropriately suggested enhancement make a clear statement whether others were already warned or not like the process to warn others was finished successfully and other people are now warned expected benefits i wrote with quite a few users on twitter via dm who were unsure if the app did warn others already or if they still have to do anything this is especially a problem if the app warned automatically in the background internal tracking id
1
2,400
5,193,008,450
IssuesEvent
2017-01-22 15:08:16
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Uncaught TypeError when doing Insert Row->Click Save. v3.0.0-beta.5
bug inprocess
ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined. This happens when I use insert row and click save. The error happens, as shown above, in ToolBar.js, line 256 when reset form is called `this.refs.form.reset();`. ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined at ToolBar.__handleSaveBtnClick__REACT_HOT_LOADER__ (ToolBar.js:256) at Object.ToolBar._this.handleSaveBtnClick [as onSave] (ToolBar.js:81) at InsertModal.__handleSave__REACT_HOT_LOADER__ (InsertModal.js:63) at InsertModal._this.handleSave (InsertModal.js:54) at InsertModalFooter.__handleSaveBtnClick__REACT_HOT_LOADER__ (InsertModalFooter.js:68) at InsertModalFooter._this.handleSaveBtnClick (InsertModalFooter.js:46) at Object.ReactErrorUtils.invokeGuardedCallback (ReactErrorUtils.js:71) at executeDispatch (EventPluginUtils.js:79) at Object.executeDispatchesInOrder (EventPluginUtils.js:102) at executeDispatchesAndRelease (EventPluginHub.js:43) at executeDispatchesAndReleaseTopLevel (EventPluginHub.js:54)
1.0
Uncaught TypeError when doing Insert Row->Click Save. v3.0.0-beta.5 - ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined. This happens when I use insert row and click save. The error happens, as shown above, in ToolBar.js, line 256 when reset form is called `this.refs.form.reset();`. ToolBar.js:256 Uncaught TypeError: Cannot read property 'reset' of undefined at ToolBar.__handleSaveBtnClick__REACT_HOT_LOADER__ (ToolBar.js:256) at Object.ToolBar._this.handleSaveBtnClick [as onSave] (ToolBar.js:81) at InsertModal.__handleSave__REACT_HOT_LOADER__ (InsertModal.js:63) at InsertModal._this.handleSave (InsertModal.js:54) at InsertModalFooter.__handleSaveBtnClick__REACT_HOT_LOADER__ (InsertModalFooter.js:68) at InsertModalFooter._this.handleSaveBtnClick (InsertModalFooter.js:46) at Object.ReactErrorUtils.invokeGuardedCallback (ReactErrorUtils.js:71) at executeDispatch (EventPluginUtils.js:79) at Object.executeDispatchesInOrder (EventPluginUtils.js:102) at executeDispatchesAndRelease (EventPluginHub.js:43) at executeDispatchesAndReleaseTopLevel (EventPluginHub.js:54)
process
uncaught typeerror when doing insert row click save beta toolbar js uncaught typeerror cannot read property reset of undefined this happens when i use insert row and click save the error happens as shown above in toolbar js line when reset form is called this refs form reset toolbar js uncaught typeerror cannot read property reset of undefined at toolbar handlesavebtnclick react hot loader toolbar js at object toolbar this handlesavebtnclick toolbar js at insertmodal handlesave react hot loader insertmodal js at insertmodal this handlesave insertmodal js at insertmodalfooter handlesavebtnclick react hot loader insertmodalfooter js at insertmodalfooter this handlesavebtnclick insertmodalfooter js at object reacterrorutils invokeguardedcallback reacterrorutils js at executedispatch eventpluginutils js at object executedispatchesinorder eventpluginutils js at executedispatchesandrelease eventpluginhub js at executedispatchesandreleasetoplevel eventpluginhub js
1
40,240
16,431,631,640
IssuesEvent
2021-05-20 02:58:18
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
Support additional parameters like hbi_workspace in azurerm_machine_learning_workspace
enhancement service/machine-learning
Some additional parameters needs to be updated as per - https://docs.microsoft.com/en-us/azure/templates/microsoft.machinelearningservices/2020-03-01/workspaces
1.0
Support additional parameters like hbi_workspace in azurerm_machine_learning_workspace - Some additional parameters needs to be updated as per - https://docs.microsoft.com/en-us/azure/templates/microsoft.machinelearningservices/2020-03-01/workspaces
non_process
support additional parameters like hbi workspace in azurerm machine learning workspace some additional parameters needs to be updated as per
0
10,669
13,457,330,206
IssuesEvent
2020-09-09 09:03:11
didi/mpx
https://api.github.com/repos/didi/mpx
closed
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面
processing
**问题描述** 使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 **环境信息描述** 1. Window 微信客户端=》小程序面板=》最近使用小程序=》点击打开小程序。 **最简复现demo** Window 微信客户端 打开滴滴出行小程序也是空白
1.0
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 - **问题描述** 使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 **环境信息描述** 1. Window 微信客户端=》小程序面板=》最近使用小程序=》点击打开小程序。 **最简复现demo** Window 微信客户端 打开滴滴出行小程序也是空白
process
使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 问题描述 使用mpx开发的小程序,在微信电脑版中打不开,都是空白页面 环境信息描述 window 微信客户端 》小程序面板 》最近使用小程序 》点击打开小程序。 最简复现demo window 微信客户端 打开滴滴出行小程序也是空白
1
362,967
25,401,239,001
IssuesEvent
2022-11-22 12:17:41
ooni/ooni.org
https://api.github.com/repos/ooni/ooni.org
opened
Documentation on how to add/update fingerprints of blocking pages
user feedback documentation priority/medium
Suggestion from the community member: > I think we need clear documentation on how to add/update fingerprints of blocking pages. It could be prominently placed somewhere on the website in a call to action section. The documentation can be added to the 'Contribute to the test lists' section and to the updated version of the 'Get involved' page.
1.0
Documentation on how to add/update fingerprints of blocking pages - Suggestion from the community member: > I think we need clear documentation on how to add/update fingerprints of blocking pages. It could be prominently placed somewhere on the website in a call to action section. The documentation can be added to the 'Contribute to the test lists' section and to the updated version of the 'Get involved' page.
non_process
documentation on how to add update fingerprints of blocking pages suggestion from the community member i think we need clear documentation on how to add update fingerprints of blocking pages it could be prominently placed somewhere on the website in a call to action section the documentation can be added to the contribute to the test lists section and to the updated version of the get involved page
0
16,428
21,300,224,012
IssuesEvent
2022-04-15 01:24:25
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
Consider swapping css-modules tooling for modular-css
:speech_balloon: RFC CSS Preprocessing ✨ Parcel 2 🔌 Parcel 2 Plugins
# 💬 RFC Proposal, swap out css-modules toolchain for modular-css. ## 🔦 Context Parcel currently has support for css-modules (yay!), but it's through very unmaintained postcss plugins from the css-modules org. These tools are pretty out of date at this point, and most of the interesting modern support for css-modules basically only exists in webpacks css-loader. Stuff, like interpolating values in selectors, (which we depend on to make component interpolation work in [astroturf](https://github.com/4Catalyzer/astroturf)) Overall the original the css-modules tooling is mostly abandoned. I've made a few attempts to try and revive/takeover/maintain these repo's but everyone if fairly unresponsive, so it seems likely they will stay that way. Since no one maintains, the pure postcss tools anymore parcel support will get more and more out of date, even if there is some life in the specific plugins that css-loader uses. All that to say, that @tivac maintains [modular-css](https://github.com/tivac/modular-css) which is functionally the same as css-modules, but with sharp edges filed down, and active development. It's not a drop in replacement, but it is very close, actively maintained, and functionally better (both to use and consume), It'd be awesome to adapt it as the basis for css module support in Parcel. it would improve the support for folks using css-modules as well as tooling like astroturf which is currently hobbled in parcel b/c of missing features. Thanks!
1.0
Consider swapping css-modules tooling for modular-css - # 💬 RFC Proposal, swap out css-modules toolchain for modular-css. ## 🔦 Context Parcel currently has support for css-modules (yay!), but it's through very unmaintained postcss plugins from the css-modules org. These tools are pretty out of date at this point, and most of the interesting modern support for css-modules basically only exists in webpacks css-loader. Stuff, like interpolating values in selectors, (which we depend on to make component interpolation work in [astroturf](https://github.com/4Catalyzer/astroturf)) Overall the original the css-modules tooling is mostly abandoned. I've made a few attempts to try and revive/takeover/maintain these repo's but everyone if fairly unresponsive, so it seems likely they will stay that way. Since no one maintains, the pure postcss tools anymore parcel support will get more and more out of date, even if there is some life in the specific plugins that css-loader uses. All that to say, that @tivac maintains [modular-css](https://github.com/tivac/modular-css) which is functionally the same as css-modules, but with sharp edges filed down, and active development. It's not a drop in replacement, but it is very close, actively maintained, and functionally better (both to use and consume), It'd be awesome to adapt it as the basis for css module support in Parcel. it would improve the support for folks using css-modules as well as tooling like astroturf which is currently hobbled in parcel b/c of missing features. Thanks!
process
consider swapping css modules tooling for modular css 💬 rfc proposal swap out css modules toolchain for modular css 🔦 context parcel currently has support for css modules yay but it s through very unmaintained postcss plugins from the css modules org these tools are pretty out of date at this point and most of the interesting modern support for css modules basically only exists in webpacks css loader stuff like interpolating values in selectors which we depend on to make component interpolation work in overall the original the css modules tooling is mostly abandoned i ve made a few attempts to try and revive takeover maintain these repo s but everyone if fairly unresponsive so it seems likely they will stay that way since no one maintains the pure postcss tools anymore parcel support will get more and more out of date even if there is some life in the specific plugins that css loader uses all that to say that tivac maintains which is functionally the same as css modules but with sharp edges filed down and active development it s not a drop in replacement but it is very close actively maintained and functionally better both to use and consume it d be awesome to adapt it as the basis for css module support in parcel it would improve the support for folks using css modules as well as tooling like astroturf which is currently hobbled in parcel b c of missing features thanks
1
254,379
8,073,429,849
IssuesEvent
2018-08-06 19:13:52
HealthCatalyst/healthcareai-r
https://api.github.com/repos/HealthCatalyst/healthcareai-r
closed
Limone integration
High Priority model interpretation new features
should be called after/separately from `predict` rather than being a switch to turn on during predict
1.0
Limone integration - should be called after/separately from `predict` rather than being a switch to turn on during predict
non_process
limone integration should be called after separately from predict rather than being a switch to turn on during predict
0
7,168
10,311,577,851
IssuesEvent
2019-08-29 17:43:01
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
PubSub: Release the first GA version
api: pubsub release blocking type: process
If no major issues are discovered during the ongoing bake-in period, the PubSub client should be released as GA. Releasing a new version is done with the help of [releasetool](https://github.com/googleapis/releasetool). Prior to the GA release, the following must be done: - Go through the checklist from the internal GA release doc, make sure everything is in order. - Changes in the client lib files: - Update PubSub Python lib README to indicate semver just like Java library has - Update beta tag in PubSub Python lib README - Update beta tag in Python libs root - Set the release level to `ga` in `.repo-metadata.json` - Change beta label in code samples in clients page in Google Cloud website - Add to release notes (talk to @kir-titievsky) - Update any documentation on Cloudsite
1.0
PubSub: Release the first GA version - If no major issues are discovered during the ongoing bake-in period, the PubSub client should be released as GA. Releasing a new version is done with the help of [releasetool](https://github.com/googleapis/releasetool). Prior to the GA release, the following must be done: - Go through the checklist from the internal GA release doc, make sure everything is in order. - Changes in the client lib files: - Update PubSub Python lib README to indicate semver just like Java library has - Update beta tag in PubSub Python lib README - Update beta tag in Python libs root - Set the release level to `ga` in `.repo-metadata.json` - Change beta label in code samples in clients page in Google Cloud website - Add to release notes (talk to @kir-titievsky) - Update any documentation on Cloudsite
process
pubsub release the first ga version if no major issues are discovered during the ongoing bake in period the pubsub client should be released as ga releasing a new version is done with the help of prior to the ga release the following must be done go through the checklist from the internal ga release doc make sure everything is in order changes in the client lib files update pubsub python lib readme to indicate semver just like java library has update beta tag in pubsub python lib readme update beta tag in python libs root set the release level to ga in repo metadata json change beta label in code samples in clients page in google cloud website add to release notes talk to kir titievsky update any documentation on cloudsite
1
337,599
24,547,309,918
IssuesEvent
2022-10-12 09:46:44
hotwax/inventory-count
https://api.github.com/repos/hotwax/inventory-count
closed
Readme file should be updated with build and run setup in hacktoberfest branch
documentation hacktoberfest
## What is the motivation for adding/enhancing this feature? <!-- Describe the motivation or the concrete use case for a new feature or why one of the current ones should be enhanced. --> Current readme file is not understandable. Make it similar as the readme file of preorder app https://github.com/hotwax/preorder of hacktoberfest branch ## What are the acceptance criteria? <!-- List the acceptance criteria for this task in the form of a list. --> - [ ] ... ## Can you complete this feature request by yourself? - [ ] YES - [ ] NO ## Additional information <!-- If you think that any additional information would be useful, please provide them here. -->
1.0
Readme file should be updated with build and run setup in hacktoberfest branch - ## What is the motivation for adding/enhancing this feature? <!-- Describe the motivation or the concrete use case for a new feature or why one of the current ones should be enhanced. --> Current readme file is not understandable. Make it similar as the readme file of preorder app https://github.com/hotwax/preorder of hacktoberfest branch ## What are the acceptance criteria? <!-- List the acceptance criteria for this task in the form of a list. --> - [ ] ... ## Can you complete this feature request by yourself? - [ ] YES - [ ] NO ## Additional information <!-- If you think that any additional information would be useful, please provide them here. -->
non_process
readme file should be updated with build and run setup in hacktoberfest branch what is the motivation for adding enhancing this feature current readme file is not understandable make it similar as the readme file of preorder app of hacktoberfest branch what are the acceptance criteria can you complete this feature request by yourself yes no additional information
0
16,762
21,934,483,325
IssuesEvent
2022-05-23 12:43:23
camunda/zeebe-process-test
https://api.github.com/repos/camunda/zeebe-process-test
opened
Test case with multiple workers is flaky
kind/bug team/process-automation
I have a strange problem in https://github.com/camunda-community-hub/spring-zeebe/ with testing where I could need some help. I currently guess it is rooted in zeebe-process-test, but are not sure. In some wired situations (which I currently haven't tracked down exactly) a job worker stops polling according to the trace log of zeebe-process-test. I can't see anything on the worker side - and I wondered if in EZE a JobAcquisition could hang for file (for whatever reason)? ![grafik](https://user-images.githubusercontent.com/1217596/169821839-50b91fe2-1298-4924-9261-41fb5fda3d9c.png) it happens when I run the test in https://github.com/berndruecker/customer-onboarding-camunda-8-springboot-extended/blob/master/src/test/java/io/berndruecker/onboarding/customer/TestCustomerOnboardingProcess.java - but not every time...
1.0
Test case with multiple workers is flaky - I have a strange problem in https://github.com/camunda-community-hub/spring-zeebe/ with testing where I could need some help. I currently guess it is rooted in zeebe-process-test, but are not sure. In some wired situations (which I currently haven't tracked down exactly) a job worker stops polling according to the trace log of zeebe-process-test. I can't see anything on the worker side - and I wondered if in EZE a JobAcquisition could hang for file (for whatever reason)? ![grafik](https://user-images.githubusercontent.com/1217596/169821839-50b91fe2-1298-4924-9261-41fb5fda3d9c.png) it happens when I run the test in https://github.com/berndruecker/customer-onboarding-camunda-8-springboot-extended/blob/master/src/test/java/io/berndruecker/onboarding/customer/TestCustomerOnboardingProcess.java - but not every time...
process
test case with multiple workers is flaky i have a strange problem in with testing where i could need some help i currently guess it is rooted in zeebe process test but are not sure in some wired situations which i currently haven t tracked down exactly a job worker stops polling according to the trace log of zeebe process test i can t see anything on the worker side and i wondered if in eze a jobacquisition could hang for file for whatever reason it happens when i run the test in but not every time
1
350
2,793,381,240
IssuesEvent
2015-05-11 10:34:54
ecodistrict/IDSSDashboard
https://api.github.com/repos/ecodistrict/IDSSDashboard
closed
Show what data is required by module
enhancement form feedback 09102014 process step: collect data
It would be good if each module would show what data it requires and what data is used by default, and also what data has been uploaded which it is using.
1.0
Show what data is required by module - It would be good if each module would show what data it requires and what data is used by default, and also what data has been uploaded which it is using.
process
show what data is required by module it would be good if each module would show what data it requires and what data is used by default and also what data has been uploaded which it is using
1
12,646
15,019,967,845
IssuesEvent
2021-02-01 14:11:02
Talendar/nevopy
https://api.github.com/repos/Talendar/nevopy
opened
Add a network processing scheduler
comp:processing type:feature
**Feature:** A processing scheduler (subclass of [`ProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/base_scheduler.py#L47)) capable of assigning work to different machines on a network. The [`RayProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/ray_processing.py#L42) seems to be able to do that to some extent, but, although I could make it work on a local network, I was unable to use it to split the processing among computers in the Internet. The network processing scheduler must be able to manage other processing schedulers (workers), each of which is in a different machine.
1.0
Add a network processing scheduler - **Feature:** A processing scheduler (subclass of [`ProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/base_scheduler.py#L47)) capable of assigning work to different machines on a network. The [`RayProcessingScheduler`](https://github.com/Talendar/nevopy/blob/e1e68921bbdb6ac73ec4bf206b3373d5b77870fd/nevopy/processing/ray_processing.py#L42) seems to be able to do that to some extent, but, although I could make it work on a local network, I was unable to use it to split the processing among computers in the Internet. The network processing scheduler must be able to manage other processing schedulers (workers), each of which is in a different machine.
process
add a network processing scheduler feature a processing scheduler subclass of capable of assigning work to different machines on a network the seems to be able to do that to some extent but although i could make it work on a local network i was unable to use it to split the processing among computers in the internet the network processing scheduler must be able to manage other processing schedulers workers each of which is in a different machine
1
845
3,314,805,819
IssuesEvent
2015-11-06 08:19:07
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
opened
Implement "select USER()"
MYSQL PROTOCOL QUERY PROCESSOR ROUTING
## WHY Some clients issue a "select USER()" . This could be executed directly from ProxySQL ## WHAT * [ ] extend MySQL_Session::handler_special_queries() to also support "select USER()"
1.0
Implement "select USER()" - ## WHY Some clients issue a "select USER()" . This could be executed directly from ProxySQL ## WHAT * [ ] extend MySQL_Session::handler_special_queries() to also support "select USER()"
process
implement select user why some clients issue a select user this could be executed directly from proxysql what extend mysql session handler special queries to also support select user
1
3,852
6,808,552,216
IssuesEvent
2017-11-04 04:30:06
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
reopened
Some command lines not working correctly
apps-makeClass status-inprocess type-bug
Some of the older tools/apps such as `makeClass`, `grabABI`, and `ethslurp` may accept -commands with a single dash instead of --command (double dash). Check all the options.cpp files for this error.
1.0
Some command lines not working correctly - Some of the older tools/apps such as `makeClass`, `grabABI`, and `ethslurp` may accept -commands with a single dash instead of --command (double dash). Check all the options.cpp files for this error.
process
some command lines not working correctly some of the older tools apps such as makeclass grababi and ethslurp may accept commands with a single dash instead of command double dash check all the options cpp files for this error
1
8,763
11,883,162,081
IssuesEvent
2020-03-27 15:31:06
prisma/prisma-client-js
https://api.github.com/repos/prisma/prisma-client-js
closed
Results should always include foreign key columns
kind/feature process/product
I would like to propose that when returning records from a table, any foreign key columns are automatically included. For example: ``` model User { id Int @id @unique name String posts Post[] } model Post { id Int @id @unique text String user User } ``` Returning all posts with `photon.posts.findMany()` today results in: ```json [ { "id": 1, "text": "Hello, world" }, { "id": 2, "text": "Foobar" } ] ``` I'm proposing that it return the following instead: ```json [ { "id": 1, "text": "Hello, world", "user": 10 }, { "id": 2, "text": "Foobar", "user": 11 } ] ``` I can still return the full associated User object, if desired, with the `include:` option. My reasoning is that when I say "I want a post" or "I want all posts" then that means I want *all* of the data for those posts. If I only want a subset of columns then I'll use the `select:` option and list the specific fields that I want. In the app I'm currently working in I'll pull a collection of data when the app launches and additional data later if the user navigates to a certain page. When they do, the additional data I need is looked up by a foreign key and then associated and displayed. I find myself constantly having to use an `:include => :select` in order to get that foreign key returned in my first batch of data. One of my queries looks like: ```javascript photon.users.findMany({ include: { note: { select: { id: true } }, responses: { include: { form: { select: { id: true } }, answers: { include: { question: { select: { id: true } } } } } } } }) ``` If the foreign keys were included by default this would become: ```javascript photon.users.findMany({ include: { note: true, responses: { include: { answers: true } } } }) ```
1.0
Results should always include foreign key columns - I would like to propose that when returning records from a table, any foreign key columns are automatically included. For example: ``` model User { id Int @id @unique name String posts Post[] } model Post { id Int @id @unique text String user User } ``` Returning all posts with `photon.posts.findMany()` today results in: ```json [ { "id": 1, "text": "Hello, world" }, { "id": 2, "text": "Foobar" } ] ``` I'm proposing that it return the following instead: ```json [ { "id": 1, "text": "Hello, world", "user": 10 }, { "id": 2, "text": "Foobar", "user": 11 } ] ``` I can still return the full associated User object, if desired, with the `include:` option. My reasoning is that when I say "I want a post" or "I want all posts" then that means I want *all* of the data for those posts. If I only want a subset of columns then I'll use the `select:` option and list the specific fields that I want. In the app I'm currently working in I'll pull a collection of data when the app launches and additional data later if the user navigates to a certain page. When they do, the additional data I need is looked up by a foreign key and then associated and displayed. I find myself constantly having to use an `:include => :select` in order to get that foreign key returned in my first batch of data. One of my queries looks like: ```javascript photon.users.findMany({ include: { note: { select: { id: true } }, responses: { include: { form: { select: { id: true } }, answers: { include: { question: { select: { id: true } } } } } } } }) ``` If the foreign keys were included by default this would become: ```javascript photon.users.findMany({ include: { note: true, responses: { include: { answers: true } } } }) ```
process
results should always include foreign key columns i would like to propose that when returning records from a table any foreign key columns are automatically included for example model user id int id unique name string posts post model post id int id unique text string user user returning all posts with photon posts findmany today results in json id text hello world id text foobar i m proposing that it return the following instead json id text hello world user id text foobar user i can still return the full associated user object if desired with the include option my reasoning is that when i say i want a post or i want all posts then that means i want all of the data for those posts if i only want a subset of columns then i ll use the select option and list the specific fields that i want in the app i m currently working in i ll pull a collection of data when the app launches and additional data later if the user navigates to a certain page when they do the additional data i need is looked up by a foreign key and then associated and displayed i find myself constantly having to use an include select in order to get that foreign key returned in my first batch of data one of my queries looks like javascript photon users findmany include note select id true responses include form select id true answers include question select id true if the foreign keys were included by default this would become javascript photon users findmany include note true responses include answers true
1
4,372
7,260,515,864
IssuesEvent
2018-02-18 10:54:36
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][processing] New algorithm for subdividing geometries
Automatic new feature Processing User Manual
Original commit: https://github.com/qgis/QGIS/commit/4dcf8d8d10249971f5ec1358a172078e11e6bad8 by nyalldawson The returned geometry will be a collection containing subdivided parts from the original geometry, where no part has more then the specified maximum number of nodes. This is useful for dividing a complex geometry into less complex parts, which are better able to be spatially indexed and faster to perform further operations such as intersects on. The returned geometry parts may not be valid and may contain self-intersections.
1.0
[FEATURE][processing] New algorithm for subdividing geometries - Original commit: https://github.com/qgis/QGIS/commit/4dcf8d8d10249971f5ec1358a172078e11e6bad8 by nyalldawson The returned geometry will be a collection containing subdivided parts from the original geometry, where no part has more then the specified maximum number of nodes. This is useful for dividing a complex geometry into less complex parts, which are better able to be spatially indexed and faster to perform further operations such as intersects on. The returned geometry parts may not be valid and may contain self-intersections.
process
new algorithm for subdividing geometries original commit by nyalldawson the returned geometry will be a collection containing subdivided parts from the original geometry where no part has more then the specified maximum number of nodes this is useful for dividing a complex geometry into less complex parts which are better able to be spatially indexed and faster to perform further operations such as intersects on the returned geometry parts may not be valid and may contain self intersections
1
80,669
23,276,156,690
IssuesEvent
2022-08-05 07:24:48
reitmas32/Next
https://api.github.com/repos/reitmas32/Next
opened
Create a basic builder
builder
## Builder that uses nothing in the base ### Example of config.yaml ```yaml basic_release: base: basic c_compiler: gcc cxx_compiler: g++ linker: ld files_cxx: - main.cpp - src/func/suma.cpp - src/structs/*.cc files_c: - main_of_c.c - src/func/suma.c - src/structs/*.c c_compiler_regex: $C $FILE -o #FILE.o cxx_compiler_regex: $CXX $FILE -o #FILE.o ld_regex: $LD $FILES -lgl -pthread ```
1.0
Create a basic builder - ## Builder that uses nothing in the base ### Example of config.yaml ```yaml basic_release: base: basic c_compiler: gcc cxx_compiler: g++ linker: ld files_cxx: - main.cpp - src/func/suma.cpp - src/structs/*.cc files_c: - main_of_c.c - src/func/suma.c - src/structs/*.c c_compiler_regex: $C $FILE -o #FILE.o cxx_compiler_regex: $CXX $FILE -o #FILE.o ld_regex: $LD $FILES -lgl -pthread ```
non_process
create a basic builder builder that uses nothing in the base example of config yaml yaml basic release base basic c compiler gcc cxx compiler g linker ld files cxx main cpp src func suma cpp src structs cc files c main of c c src func suma c src structs c c compiler regex c file o file o cxx compiler regex cxx file o file o ld regex ld files lgl pthread
0
18,109
24,135,657,568
IssuesEvent
2022-09-21 11:03:50
python/cpython
https://api.github.com/repos/python/cpython
closed
ProcessPoolExecutor: float32 result of executors seems to get truncated.
type-bug pending expert-multiprocessing
# Bug report ``` import time import numpy as np from concurrent.futures import ProcessPoolExecutor def example_executor(): time.sleep(1) return np.float32(103167328256.2) if __name__ == '__main__': processing_pool = ProcessPoolExecutor() future = processing_pool.submit(example_executor) while not future.done(): time.sleep(1) print(future.result()) ``` yields: ` 103167330000.0 ` # Your environment Windows 10. Python 3.10.2 and also 3.11.0rc2
1.0
ProcessPoolExecutor: float32 result of executors seems to get truncated. - # Bug report ``` import time import numpy as np from concurrent.futures import ProcessPoolExecutor def example_executor(): time.sleep(1) return np.float32(103167328256.2) if __name__ == '__main__': processing_pool = ProcessPoolExecutor() future = processing_pool.submit(example_executor) while not future.done(): time.sleep(1) print(future.result()) ``` yields: ` 103167330000.0 ` # Your environment Windows 10. Python 3.10.2 and also 3.11.0rc2
process
processpoolexecutor result of executors seems to get truncated bug report import time import numpy as np from concurrent futures import processpoolexecutor def example executor time sleep return np if name main processing pool processpoolexecutor future processing pool submit example executor while not future done time sleep print future result yields your environment windows python and also
1
11,319
14,138,102,894
IssuesEvent
2020-11-10 07:53:16
didi/mpx
https://api.github.com/repos/didi/mpx
closed
QQ小程序和微信小程序的订阅消息接口差异
processing
# 描述 在 QQ小程序 中订阅消息的接口为 `qq.subscribeAppMsg` [文档](https://q.qq.com/wiki/develop/miniprogram/API/open_port/port_subscription.html) 在微信小程序中订阅消息调用为 `wx.requestSubscribeMessage` [文档](https://developers.weixin.qq.com/miniprogram/dev/api/open-api/subscribe-message/wx.requestSubscribeMessage.html) 在使用 mpx 推荐的兼容写法 `mpx.requestSubscribeMessage` 无法实现自动代理到 QQ小程序 对应的API中。对于这种情况下 mpx 目前的方案可以通过配置实现吗? # 备注 可以在 QQ小程序 通过 `mpx.subscribeAppMsg` 调用到接,但是这样就需要做平台判断。另外,由于项目使用了 TypeScript ,这样使用会在类型推导上产生错误。
1.0
QQ小程序和微信小程序的订阅消息接口差异 - # 描述 在 QQ小程序 中订阅消息的接口为 `qq.subscribeAppMsg` [文档](https://q.qq.com/wiki/develop/miniprogram/API/open_port/port_subscription.html) 在微信小程序中订阅消息调用为 `wx.requestSubscribeMessage` [文档](https://developers.weixin.qq.com/miniprogram/dev/api/open-api/subscribe-message/wx.requestSubscribeMessage.html) 在使用 mpx 推荐的兼容写法 `mpx.requestSubscribeMessage` 无法实现自动代理到 QQ小程序 对应的API中。对于这种情况下 mpx 目前的方案可以通过配置实现吗? # 备注 可以在 QQ小程序 通过 `mpx.subscribeAppMsg` 调用到接,但是这样就需要做平台判断。另外,由于项目使用了 TypeScript ,这样使用会在类型推导上产生错误。
process
qq小程序和微信小程序的订阅消息接口差异 描述 在 qq小程序 中订阅消息的接口为 qq subscribeappmsg 在微信小程序中订阅消息调用为 wx requestsubscribemessage 在使用 mpx 推荐的兼容写法 mpx requestsubscribemessage 无法实现自动代理到 qq小程序 对应的api中。对于这种情况下 mpx 目前的方案可以通过配置实现吗? 备注 可以在 qq小程序 通过 mpx subscribeappmsg 调用到接,但是这样就需要做平台判断。另外,由于项目使用了 typescript ,这样使用会在类型推导上产生错误。
1
10,190
13,045,131,038
IssuesEvent
2020-07-29 06:35:01
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
The Start/Stop VMs solution during off-hours: bug with Resource Group naming
Pri2 automation/svc cxp process-automation/subsvc product-question triaged
[Enter feedback here] It seems that the automation is bugging with resource group with a . in the name Error message XXX.Production is not a valid ResourceGroup Name. Please verify your input! Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Azure Automation Start/Stop VMs during off-hours overview](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
The Start/Stop VMs solution during off-hours: bug with Resource Group naming - [Enter feedback here] It seems that the automation is bugging with resource group with a . in the name Error message XXX.Production is not a valid ResourceGroup Name. Please verify your input! Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Azure Automation Start/Stop VMs during off-hours overview](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
the start stop vms solution during off hours bug with resource group naming it seems that the automation is bugging with resource group with a in the name error message xxx production is not a valid resourcegroup name please verify your input thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
157,566
13,696,986,465
IssuesEvent
2020-10-01 01:40:45
gurupratap-matharu/volu
https://api.github.com/repos/gurupratap-matharu/volu
opened
Inherit base template to base_with_header and base_with_main html templates
documentation enhancement
- Use _base.html to create base_with_header html template which has a nice page header - Similarly, use _base.html to crate a base_with_main html template which has access to directy main content in child templates This way we do not have to maintain margin and paddings in css in all templates Use simple template inheritance and our view templates should directly write content in specified containers without being bothered about responsiveness and alignments
1.0
Inherit base template to base_with_header and base_with_main html templates - - Use _base.html to create base_with_header html template which has a nice page header - Similarly, use _base.html to crate a base_with_main html template which has access to directy main content in child templates This way we do not have to maintain margin and paddings in css in all templates Use simple template inheritance and our view templates should directly write content in specified containers without being bothered about responsiveness and alignments
non_process
inherit base template to base with header and base with main html templates use base html to create base with header html template which has a nice page header similarly use base html to crate a base with main html template which has access to directy main content in child templates this way we do not have to maintain margin and paddings in css in all templates use simple template inheritance and our view templates should directly write content in specified containers without being bothered about responsiveness and alignments
0
163,671
25,856,163,682
IssuesEvent
2022-12-13 13:55:36
decidim/decidim
https://api.github.com/repos/decidim/decidim
opened
HERE Maps static map size
contract: redesign
If you use HERE maps, currently, a small static map is generated: <img width="811" alt="CleanShot 2022-12-13 at 14 49 54@2x" src="https://user-images.githubusercontent.com/24740/207345569-b9b47136-0a3a-41e4-9d13-34dd201037cb.png"> (https://www.decidim.barcelona/assemblies/barribesosmaresme/f/3229/meetings/5765) In the new design, this static map is slightly bigger but as we are using the same call, the map is pixelated: <img width="1001" alt="CleanShot 2022-12-13 at 14 54 07@2x" src="https://user-images.githubusercontent.com/24740/207347884-283fcf07-5018-4843-bfd5-bf467913b7a1.png"> The call to the map should be changed (or made dynamic?) to adapt to the new size. https://github.com/decidim/decidim/blob/db91cf387557bfccee917e744303b25d06c691c4/decidim-generators/lib/decidim/generators/app_templates/initializer.rb#L49 https://developer.here.com/documentation/map-image/dev_guide/topics/display-options.html
1.0
HERE Maps static map size - If you use HERE maps, currently, a small static map is generated: <img width="811" alt="CleanShot 2022-12-13 at 14 49 54@2x" src="https://user-images.githubusercontent.com/24740/207345569-b9b47136-0a3a-41e4-9d13-34dd201037cb.png"> (https://www.decidim.barcelona/assemblies/barribesosmaresme/f/3229/meetings/5765) In the new design, this static map is slightly bigger but as we are using the same call, the map is pixelated: <img width="1001" alt="CleanShot 2022-12-13 at 14 54 07@2x" src="https://user-images.githubusercontent.com/24740/207347884-283fcf07-5018-4843-bfd5-bf467913b7a1.png"> The call to the map should be changed (or made dynamic?) to adapt to the new size. https://github.com/decidim/decidim/blob/db91cf387557bfccee917e744303b25d06c691c4/decidim-generators/lib/decidim/generators/app_templates/initializer.rb#L49 https://developer.here.com/documentation/map-image/dev_guide/topics/display-options.html
non_process
here maps static map size if you use here maps currently a small static map is generated img width alt cleanshot at src in the new design this static map is slightly bigger but as we are using the same call the map is pixelated img width alt cleanshot at src the call to the map should be changed or made dynamic to adapt to the new size
0
21,393
29,202,232,295
IssuesEvent
2023-05-21 00:37:44
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] Data Scientist na Coodesh
SALVADOR BANCO DE DADOS BIG DATA DATA SCIENCE SQL AWS REQUISITOS REMOTO PROCESSOS INOVAÇÃO GITHUB CI UMA QUALIDADE R BI MINERAÇÃO DE DADOS MACHINE LEARNING NEGÓCIOS IOT Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>First Decision</strong> está em busca de <strong><ins>Data Scientist</ins></strong><strong> </strong>para compor seu time!</p> <p>A First Decision é uma empresa de tecnologia com expertise em BI, Data Governance e Data Quality. Além disso, oferecemos soluções em tecnologias como IoT, Machine Learning, Big Data, Biometria, entre outras. Somos parceiros Hitachi Vantara, AWS e Mongo DB. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos parceiros auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p> <p></p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Analisar, modelar, propor, implementar e avaliar soluções para problemas complexos e simples de negócios, por meio de técnicas estatísticas, matemáticas e computacionais aplicadas ao processamento e visualização de dados.</li> </ul> ## First Decision: <p>Trazemos vantagem competitiva e auxiliamos nas tomadas de decisões para que sejam as mais inteligentes e vantajosas possíveis. Composta de colaboradores altamente capacitados e especializados, estamos sempre pesquisando e idealizando novas soluções que permitem aos nossos clientes gerenciar os dados de suas organizações para impulsionar a eficiência operacional. Através da criação e implementação de soluções que aprimorem a maneira de relacionar o seu negócio com a sociedade. Conduzimos os processos de transformação através da inovação para potencializar seus resultados. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos clientes auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p> </p> ## Habilidades: - Big Data Architects - Banco de dados relacionais (SQL) - Machine Learning ## Local: 100% Remoto ## Requisitos: - Experiência sólida com Big Data; - Ampla experiência em Ciência de Dados; - Superior completo em TI; - Domínio da linguagem SQL; - Domínio de SGBDs; - Conhecimento avançado em mineração de dados, extração de informações não triviais e aprendizagem de máquina. ## Benefícios: - Vale alimentação: R$ 32,17 por dia; - Plano de Saúde: 50% custeada pela empresa; - Plano Odontológico; - Plataforma para terapia; - Seguro de vida. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Data Scientist na First Decision](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime CLT #### Categoria Data Science
1.0
[Remoto] Data Scientist na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>First Decision</strong> está em busca de <strong><ins>Data Scientist</ins></strong><strong> </strong>para compor seu time!</p> <p>A First Decision é uma empresa de tecnologia com expertise em BI, Data Governance e Data Quality. Além disso, oferecemos soluções em tecnologias como IoT, Machine Learning, Big Data, Biometria, entre outras. Somos parceiros Hitachi Vantara, AWS e Mongo DB. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos parceiros auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p> <p></p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Analisar, modelar, propor, implementar e avaliar soluções para problemas complexos e simples de negócios, por meio de técnicas estatísticas, matemáticas e computacionais aplicadas ao processamento e visualização de dados.</li> </ul> ## First Decision: <p>Trazemos vantagem competitiva e auxiliamos nas tomadas de decisões para que sejam as mais inteligentes e vantajosas possíveis. Composta de colaboradores altamente capacitados e especializados, estamos sempre pesquisando e idealizando novas soluções que permitem aos nossos clientes gerenciar os dados de suas organizações para impulsionar a eficiência operacional. Através da criação e implementação de soluções que aprimorem a maneira de relacionar o seu negócio com a sociedade. Conduzimos os processos de transformação através da inovação para potencializar seus resultados. Tudo isso faz da First Decision uma empresa moderna e à frente do seu tempo, que oferece soluções inteligentes, condiciona soluções integradas e agrega valores à gestão de negócios de nossos clientes auxiliando na sua competitividade. Sempre visando o desenvolvimento sustentável, inovação e qualidade, pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor.</p> </p> ## Habilidades: - Big Data Architects - Banco de dados relacionais (SQL) - Machine Learning ## Local: 100% Remoto ## Requisitos: - Experiência sólida com Big Data; - Ampla experiência em Ciência de Dados; - Superior completo em TI; - Domínio da linguagem SQL; - Domínio de SGBDs; - Conhecimento avançado em mineração de dados, extração de informações não triviais e aprendizagem de máquina. ## Benefícios: - Vale alimentação: R$ 32,17 por dia; - Plano de Saúde: 50% custeada pela empresa; - Plano Odontológico; - Plataforma para terapia; - Seguro de vida. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Data Scientist na First Decision](https://coodesh.com/vagas/cientista-de-dados-132926788?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime CLT #### Categoria Data Science
process
data scientist na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a first decision está em busca de data scientist para compor seu time a first decision é uma empresa de tecnologia com expertise em bi data governance e data quality além disso oferecemos soluções em tecnologias como iot machine learning big data biometria entre outras somos parceiros hitachi vantara aws e mongo db tudo isso faz da first decision uma empresa moderna e à frente do seu tempo que oferece soluções inteligentes condiciona soluções integradas e agrega valores à gestão de negócios de nossos parceiros auxiliando na sua competitividade sempre visando o desenvolvimento sustentável inovação e qualidade pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor responsabilidades analisar modelar propor implementar e avaliar soluções para problemas complexos e simples de negócios por meio de técnicas estatísticas matemáticas e computacionais aplicadas ao processamento e visualização de dados first decision trazemos vantagem competitiva e auxiliamos nas tomadas de decisões para que sejam as mais inteligentes e vantajosas possíveis composta de colaboradores altamente capacitados e especializados estamos sempre pesquisando e idealizando novas soluções que permitem aos nossos clientes gerenciar os dados de suas organizações para impulsionar a eficiência operacional através da criação e implementação de soluções que aprimorem a maneira de relacionar o seu negócio com a sociedade conduzimos os processos de transformação através da inovação para potencializar seus resultados tudo isso faz da first decision uma empresa moderna e à frente do seu tempo que oferece soluções inteligentes condiciona soluções integradas e agrega valores à gestão de negócios de nossos clientes auxiliando na sua competitividade sempre visando o desenvolvimento sustentável inovação e qualidade pois sabemos que as soluções são ainda mais valiosas quando dirigida para uma vida e um futuro melhor habilidades big data architects banco de dados relacionais sql machine learning local remoto requisitos experiência sólida com big data ampla experiência em ciência de dados superior completo em ti domínio da linguagem sql domínio de sgbds conhecimento avançado em mineração de dados extração de informações não triviais e aprendizagem de máquina benefícios vale alimentação r por dia plano de saúde custeada pela empresa plano odontológico plataforma para terapia seguro de vida como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime clt categoria data science
1
333,941
24,397,618,737
IssuesEvent
2022-10-04 20:51:20
Leafwing-Studios/leafwing_input_playback
https://api.github.com/repos/Leafwing-Studios/leafwing_input_playback
closed
Add basic instructions to README
documentation
We should explain a bit more about how this crate is intended to do, how it works and so on.
1.0
Add basic instructions to README - We should explain a bit more about how this crate is intended to do, how it works and so on.
non_process
add basic instructions to readme we should explain a bit more about how this crate is intended to do how it works and so on
0
112,295
17,089,398,484
IssuesEvent
2021-07-08 15:30:20
keep-network/coverage-pools
https://api.github.com/repos/keep-network/coverage-pools
closed
Flashloan can be used to increase rewards when calling the RiskManager
:eyeglasses: security-audit
``` Severity: Medium Difficulty: Medium Type: Data Validation Target: RiskManager.sol, CoveragePool.sol ``` The Risk Manager rewards users that report liquidations, but they can manipulate the value received using flashloans. `covAmountToGrant` determines the reward to grant the uses that notify when deposits are in liquidations or liquidated using `RiskManager`: https://github.com/keep-network/coverage-pools/blob/main/contracts/RiskManagerV1.sol#L909-L943 The computation of the reward depends directly on the total supply of `underwriterToken`: https://github.com/keep-network/coverage-pools/blob/main/contracts/CoveragePool.sol#L231-L241 A user can temporarily increase this value by making a series of calls to do the following in a single transaction: 1. Take out a flash loan 2. Make a large deposit in the underwriter token 3. Notify a liquidation and receive the reward 4. Sell the underwriter tokens in a secondary market 5. Repay the flash loan **Exploit Scenario** Eve wants to profit from the RiskManager so she takes a flashloan to increase her reward when reporting liquidations and is able to manipulate the system in her favor. **Recommendations** Short term, revise the reward computation to avoid using `assetPool.underwriterToken().totalSupply()`, which depends on the total supply of the underwriter token. Long term, review the specification on incentives and penalties to make sure that they do not depend on variables that can be manipulated by users. Use Echidna or Manticore to test them.
True
Flashloan can be used to increase rewards when calling the RiskManager - ``` Severity: Medium Difficulty: Medium Type: Data Validation Target: RiskManager.sol, CoveragePool.sol ``` The Risk Manager rewards users that report liquidations, but they can manipulate the value received using flashloans. `covAmountToGrant` determines the reward to grant the uses that notify when deposits are in liquidations or liquidated using `RiskManager`: https://github.com/keep-network/coverage-pools/blob/main/contracts/RiskManagerV1.sol#L909-L943 The computation of the reward depends directly on the total supply of `underwriterToken`: https://github.com/keep-network/coverage-pools/blob/main/contracts/CoveragePool.sol#L231-L241 A user can temporarily increase this value by making a series of calls to do the following in a single transaction: 1. Take out a flash loan 2. Make a large deposit in the underwriter token 3. Notify a liquidation and receive the reward 4. Sell the underwriter tokens in a secondary market 5. Repay the flash loan **Exploit Scenario** Eve wants to profit from the RiskManager so she takes a flashloan to increase her reward when reporting liquidations and is able to manipulate the system in her favor. **Recommendations** Short term, revise the reward computation to avoid using `assetPool.underwriterToken().totalSupply()`, which depends on the total supply of the underwriter token. Long term, review the specification on incentives and penalties to make sure that they do not depend on variables that can be manipulated by users. Use Echidna or Manticore to test them.
non_process
flashloan can be used to increase rewards when calling the riskmanager severity medium difficulty medium type data validation target riskmanager sol coveragepool sol the risk manager rewards users that report liquidations but they can manipulate the value received using flashloans covamounttogrant determines the reward to grant the uses that notify when deposits are in liquidations or liquidated using riskmanager the computation of the reward depends directly on the total supply of underwritertoken a user can temporarily increase this value by making a series of calls to do the following in a single transaction take out a flash loan make a large deposit in the underwriter token notify a liquidation and receive the reward sell the underwriter tokens in a secondary market repay the flash loan exploit scenario eve wants to profit from the riskmanager so she takes a flashloan to increase her reward when reporting liquidations and is able to manipulate the system in her favor recommendations short term revise the reward computation to avoid using assetpool underwritertoken totalsupply which depends on the total supply of the underwriter token long term review the specification on incentives and penalties to make sure that they do not depend on variables that can be manipulated by users use echidna or manticore to test them
0
186,765
21,969,349,641
IssuesEvent
2022-05-25 01:13:30
DarshiL7/Authentication-demo-Project
https://api.github.com/repos/DarshiL7/Authentication-demo-Project
opened
Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll: 1 vulnerabilities (highest severity is: 5.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p></summary> <p>Microsoft.AspNetCore.Authentication.JwtBearer</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p> <p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-34532](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll | Direct | Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-34532</summary> ### Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p> <p>Microsoft.AspNetCore.Authentication.JwtBearer</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p> <p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p> <p> Dependency Hierarchy: - :x: **Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> ASP.NET Core and Visual Studio Information Disclosure Vulnerability <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532>CVE-2021-34532</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q7cg-43mg-qp69">https://github.com/advisories/GHSA-q7cg-43mg-qp69</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"Microsoft.AspNetCore.Authentication.JwtBearer","packageVersion":"3.100.1020.52004","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"Microsoft.AspNetCore.Authentication.JwtBearer:3.100.1020.52004","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-34532","vulnerabilityDetails":"ASP.NET Core and Visual Studio Information Disclosure Vulnerability","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}]</REMEDIATE> -->
True
Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll: 1 vulnerabilities (highest severity is: 5.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p></summary> <p>Microsoft.AspNetCore.Authentication.JwtBearer</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p> <p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2021-34532](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll | Direct | Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-34532</summary> ### Vulnerable Library - <b>Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll</b></p> <p>Microsoft.AspNetCore.Authentication.JwtBearer</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg">https://api.nuget.org/packages/microsoft.aspnetcore.authentication.jwtbearer.3.1.10.nupkg</a></p> <p>Path to vulnerable library: /Authentication-Demo-Project.Api/bin/Debug/netcoreapp3.1/Microsoft.AspNetCore.Authentication.JwtBearer.dll</p> <p> Dependency Hierarchy: - :x: **Microsoft.AspNetCore.Authentication.JwtBearer-3.100.1020.52004.dll** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DarshiL7/Authentication-demo-Project/commit/58e50b4be59ccbd111c28f69520357815e88ab69">58e50b4be59ccbd111c28f69520357815e88ab69</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> ASP.NET Core and Visual Studio Information Disclosure Vulnerability <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532>CVE-2021-34532</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q7cg-43mg-qp69">https://github.com/advisories/GHSA-q7cg-43mg-qp69</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"Microsoft.AspNetCore.Authentication.JwtBearer","packageVersion":"3.100.1020.52004","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"Microsoft.AspNetCore.Authentication.JwtBearer:3.100.1020.52004","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.AspNetCore.Authentication.JwtBearer - 2.1.30, 3.1.18, 5.0.9","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-34532","vulnerabilityDetails":"ASP.NET Core and Visual Studio Information Disclosure Vulnerability","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-34532","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}]</REMEDIATE> -->
non_process
microsoft aspnetcore authentication jwtbearer dll vulnerabilities highest severity is vulnerable library microsoft aspnetcore authentication jwtbearer dll microsoft aspnetcore authentication jwtbearer library home page a href path to vulnerable library authentication demo project api bin debug microsoft aspnetcore authentication jwtbearer dll found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium microsoft aspnetcore authentication jwtbearer dll direct microsoft aspnetcore authentication jwtbearer details cve vulnerable library microsoft aspnetcore authentication jwtbearer dll microsoft aspnetcore authentication jwtbearer library home page a href path to vulnerable library authentication demo project api bin debug microsoft aspnetcore authentication jwtbearer dll dependency hierarchy x microsoft aspnetcore authentication jwtbearer dll vulnerable library found in head commit a href found in base branch main vulnerability details asp net core and visual studio information disclosure vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore authentication jwtbearer step up your open source security game with whitesource istransitivedependency false dependencytree microsoft aspnetcore authentication jwtbearer isminimumfixversionavailable true minimumfixversion microsoft aspnetcore authentication jwtbearer isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails asp net core and visual studio information disclosure vulnerability vulnerabilityurl
0
19,133
25,187,442,620
IssuesEvent
2022-11-11 19:35:19
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
opened
Páginas extras coletadas ao coletar arquivos através de navegação do browser
[1] Bug [0] Desenvolvimento [2] Média Prioridade [3] Processamento Dinâmico
## Comportamento Esperado Quando clicamos em um link (ou qualquer ação similar) que leve ao download de um arquivo no mecanismo de passos, desejamos que o arquivo seja baixado sem gerar nenhuma página correspondente. ## Comportamento Atual Ao gerar um download de arquivo através de navegação no mecanismo de passos, o arquivo é baixado corretamente, mas uma página vazia correspondente também é criada. ## Passos para reproduzir o erro Executar a coleta disponibilizada abaixo, e verificar que na página temos 8 links, dos quais 7 levam a outras páginas e 1 leva a um arquivo. Após a execução, verificar que o arquivo é baixado corretamente, mas que 9 páginas foram salvas. O esperado seriam 8 (a página inicial + os 7 links). É possivel identificar que uma das páginas está vazia, e portanto deve corresponder ao arquivo baixado. ## Especificações da Coleta ``` { "source_name": "[Realiza\u00e7\u00e3o-F01] Coleta de Informa\u00e7\u00f5es_Intitucionais_do_Munic\u00edpio - Leis_Municipais de Santana dos Montes", "base_url": "http:\/\/pt.santanadosmontes.mg.gov.br\/Legislacao", "obey_robots": false, "ignore_data_crawled_in_previous_instances": false, "crawler_description": "a", "crawler_type_desc": "Contratos", "crawler_issue": 0, "data_path": "santana_dos_montes_leis_municipais", "sc_scheduler_persist": true, "sc_scheduler_queue_refresh": 10, "sc_queue_hits": 10, "sc_queue_window": 60, "sc_queue_moderated": true, "sc_dupefilter_timeout": 600, "sc_global_page_per_domain_limit": null, "sc_global_page_per_domain_limit_timeout": 600, "sc_domain_max_page_timeout": 600, "sc_scheduler_ip_refresh": 60, "sc_scheduler_backlog_blacklist": true, "sc_scheduler_type_enabled": true, "sc_scheduler_ip_enabled": true, "sc_scheduler_item_retries": 3, "sc_scheduler_queue_timeout": 3600, "sc_httperror_allow_all": true, "sc_retry_times": 3, "sc_download_timeout": 10, "antiblock_download_delay": 2, "antiblock_autothrottle_enabled": false, "antiblock_autothrottle_start_delay": 2, "antiblock_autothrottle_max_delay": 10, "antiblock_ip_rotation_enabled": false, "antiblock_ip_rotation_type": "tor", "antiblock_max_reqs_per_ip": 10, "antiblock_max_reuse_rounds": 10, "antiblock_proxy_list": "", "antiblock_user_agent_rotation_enabled": false, "antiblock_reqs_per_user_agent": 100, "antiblock_user_agents_list": "", "antiblock_insert_cookies_enabled": false, "antiblock_cookies_list": "", "captcha": "none", "has_webdriver": false, "webdriver_path": null, "img_xpath": null, "sound_xpath": null, "dynamic_processing": true, "browser_type": "chromium", "skip_iter_errors": true, "browser_resolution_width": 1280, "browser_resolution_height": 720, "explore_links": false, "link_extractor_max_depth": 1, "link_extractor_allow_url": null, "link_extractor_allow_domains": null, "link_extractor_tags": null, "link_extractor_attrs": null, "link_extractor_check_type": false, "link_extractor_process_value": "", "download_files": false, "download_files_allow_url": null, "download_files_allow_extensions": ".pdf", "download_files_allow_domains": null, "download_files_tags": null, "download_files_attrs": null, "download_files_process_value": "", "download_files_check_large_content": true, "download_imgs": false, "steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"salva_pagina\",\"depth\":1,\"arguments\":{}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"link\",\"children\":[{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"link\"}},{\"step\":\"abrir_em_nova_aba\",\"depth\":2,\"link_xpath\":\"link\",\"children\":[]},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"salva_pagina\",\"depth\":2,\"arguments\":{}},{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"\\\"p\u00e1gina salva\\\"\"}},{\"step\":\"fechar_aba\",\"depth\":2,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"\\\"\/\/tr[*]\/td\/span\/a\\\"\"}}}}]}", "encoding_detection_method": 1, "expected_runtime_category": "medium", "templated_url_parameter_handlers": [], "templated_url_response_handlers": [], "instance_id": "166819348934970", "crawler_id": 52 } ```
1.0
Páginas extras coletadas ao coletar arquivos através de navegação do browser - ## Comportamento Esperado Quando clicamos em um link (ou qualquer ação similar) que leve ao download de um arquivo no mecanismo de passos, desejamos que o arquivo seja baixado sem gerar nenhuma página correspondente. ## Comportamento Atual Ao gerar um download de arquivo através de navegação no mecanismo de passos, o arquivo é baixado corretamente, mas uma página vazia correspondente também é criada. ## Passos para reproduzir o erro Executar a coleta disponibilizada abaixo, e verificar que na página temos 8 links, dos quais 7 levam a outras páginas e 1 leva a um arquivo. Após a execução, verificar que o arquivo é baixado corretamente, mas que 9 páginas foram salvas. O esperado seriam 8 (a página inicial + os 7 links). É possivel identificar que uma das páginas está vazia, e portanto deve corresponder ao arquivo baixado. ## Especificações da Coleta ``` { "source_name": "[Realiza\u00e7\u00e3o-F01] Coleta de Informa\u00e7\u00f5es_Intitucionais_do_Munic\u00edpio - Leis_Municipais de Santana dos Montes", "base_url": "http:\/\/pt.santanadosmontes.mg.gov.br\/Legislacao", "obey_robots": false, "ignore_data_crawled_in_previous_instances": false, "crawler_description": "a", "crawler_type_desc": "Contratos", "crawler_issue": 0, "data_path": "santana_dos_montes_leis_municipais", "sc_scheduler_persist": true, "sc_scheduler_queue_refresh": 10, "sc_queue_hits": 10, "sc_queue_window": 60, "sc_queue_moderated": true, "sc_dupefilter_timeout": 600, "sc_global_page_per_domain_limit": null, "sc_global_page_per_domain_limit_timeout": 600, "sc_domain_max_page_timeout": 600, "sc_scheduler_ip_refresh": 60, "sc_scheduler_backlog_blacklist": true, "sc_scheduler_type_enabled": true, "sc_scheduler_ip_enabled": true, "sc_scheduler_item_retries": 3, "sc_scheduler_queue_timeout": 3600, "sc_httperror_allow_all": true, "sc_retry_times": 3, "sc_download_timeout": 10, "antiblock_download_delay": 2, "antiblock_autothrottle_enabled": false, "antiblock_autothrottle_start_delay": 2, "antiblock_autothrottle_max_delay": 10, "antiblock_ip_rotation_enabled": false, "antiblock_ip_rotation_type": "tor", "antiblock_max_reqs_per_ip": 10, "antiblock_max_reuse_rounds": 10, "antiblock_proxy_list": "", "antiblock_user_agent_rotation_enabled": false, "antiblock_reqs_per_user_agent": 100, "antiblock_user_agents_list": "", "antiblock_insert_cookies_enabled": false, "antiblock_cookies_list": "", "captcha": "none", "has_webdriver": false, "webdriver_path": null, "img_xpath": null, "sound_xpath": null, "dynamic_processing": true, "browser_type": "chromium", "skip_iter_errors": true, "browser_resolution_width": 1280, "browser_resolution_height": 720, "explore_links": false, "link_extractor_max_depth": 1, "link_extractor_allow_url": null, "link_extractor_allow_domains": null, "link_extractor_tags": null, "link_extractor_attrs": null, "link_extractor_check_type": false, "link_extractor_process_value": "", "download_files": false, "download_files_allow_url": null, "download_files_allow_extensions": ".pdf", "download_files_allow_domains": null, "download_files_tags": null, "download_files_attrs": null, "download_files_process_value": "", "download_files_check_large_content": true, "download_imgs": false, "steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"screenshot\",\"depth\":1,\"arguments\":{}},{\"step\":\"salva_pagina\",\"depth\":1,\"arguments\":{}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"link\",\"children\":[{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"link\"}},{\"step\":\"abrir_em_nova_aba\",\"depth\":2,\"link_xpath\":\"link\",\"children\":[]},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"5\"}},{\"step\":\"salva_pagina\",\"depth\":2,\"arguments\":{}},{\"step\":\"imprime\",\"depth\":2,\"arguments\":{\"texto\":\"\\\"p\u00e1gina salva\\\"\"}},{\"step\":\"fechar_aba\",\"depth\":2,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"\\\"\/\/tr[*]\/td\/span\/a\\\"\"}}}}]}", "encoding_detection_method": 1, "expected_runtime_category": "medium", "templated_url_parameter_handlers": [], "templated_url_response_handlers": [], "instance_id": "166819348934970", "crawler_id": 52 } ```
process
páginas extras coletadas ao coletar arquivos através de navegação do browser comportamento esperado quando clicamos em um link ou qualquer ação similar que leve ao download de um arquivo no mecanismo de passos desejamos que o arquivo seja baixado sem gerar nenhuma página correspondente comportamento atual ao gerar um download de arquivo através de navegação no mecanismo de passos o arquivo é baixado corretamente mas uma página vazia correspondente também é criada passos para reproduzir o erro executar a coleta disponibilizada abaixo e verificar que na página temos links dos quais levam a outras páginas e leva a um arquivo após a execução verificar que o arquivo é baixado corretamente mas que páginas foram salvas o esperado seriam a página inicial os links é possivel identificar que uma das páginas está vazia e portanto deve corresponder ao arquivo baixado especificações da coleta source name coleta de informa intitucionais do munic leis municipais de santana dos montes base url http pt santanadosmontes mg gov br legislacao obey robots false ignore data crawled in previous instances false crawler description a crawler type desc contratos crawler issue data path santana dos montes leis municipais sc scheduler persist true sc scheduler queue refresh sc queue hits sc queue window sc queue moderated true sc dupefilter timeout sc global page per domain limit null sc global page per domain limit timeout sc domain max page timeout sc scheduler ip refresh sc scheduler backlog blacklist true sc scheduler type enabled true sc scheduler ip enabled true sc scheduler item retries sc scheduler queue timeout sc httperror allow all true sc retry times sc download timeout antiblock download delay antiblock autothrottle enabled false antiblock autothrottle start delay antiblock autothrottle max delay antiblock ip rotation enabled false antiblock ip rotation type tor antiblock max reqs per ip antiblock max reuse rounds antiblock proxy list antiblock user agent rotation enabled false antiblock reqs per user agent antiblock user agents list antiblock insert cookies enabled false antiblock cookies list captcha none has webdriver false webdriver path null img xpath null sound xpath null dynamic processing true browser type chromium skip iter errors true browser resolution width browser resolution height explore links false link extractor max depth link extractor allow url null link extractor allow domains null link extractor tags null link extractor attrs null link extractor check type false link extractor process value download files false download files allow url null download files allow extensions pdf download files allow domains null download files tags null download files attrs null download files process value download files check large content true download imgs false steps step root depth children step espere depth arguments segundos step salva pagina depth arguments step imprime depth arguments texto p salva step fechar aba depth arguments iterable call step localiza elementos arguments xpath tr td span a encoding detection method expected runtime category medium templated url parameter handlers templated url response handlers instance id crawler id
1
1,978
4,805,290,972
IssuesEvent
2016-11-02 15:42:11
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
A bug displaying boolean value
bug inprocess
Hi, I had some data like [{test:true},{test:true},{test:false}] The True value could display well, but the false value was empty. Could you check this? Best, Yang
1.0
A bug displaying boolean value - Hi, I had some data like [{test:true},{test:true},{test:false}] The True value could display well, but the false value was empty. Could you check this? Best, Yang
process
a bug displaying boolean value hi i had some data like the true value could display well but the false value was empty could you check this best yang
1
100,546
30,723,069,200
IssuesEvent
2023-07-27 17:22:47
spack/spack
https://api.github.com/repos/spack/spack
opened
Installation issue: cgns tries to use mpifc even if fortran is disabled
build-error
### Steps to reproduce the issue ```console $ spack spec -I cgns@4.3.0~fortran ^mpich@4.1.1~fortran Input spec -------------------------------- - cgns@4.3.0~fortran - ^mpich@4.1.1~fortran Concretized -------------------------------- - cgns@4.3.0%clang@16.0.0~base_scope~fortran+hdf5~int64~ipo~legacy~mem_debug+mpi+scoping+shared~static~testing~tools build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell [+] ^cmake@3.25.1%clang@16.0.0~doc+ncurses+ownlibs~qt build_system=generic build_type=Release arch=linux-ubuntu20.04-haswell [+] ^gmake@4.2.1%clang@16.0.0~guile build_system=autotools patches=ca60bd9,fe5b60d arch=linux-ubuntu20.04-haswell [+] ^hdf5@1.12.2%clang@16.0.0~cxx~fortran~hl~ipo~java+mpi+shared~szip~threadsafe+tools api=default build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell [+] ^pkgconf@1.8.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^zlib@1.2.13%clang@16.0.0+optimize+pic+shared build_system=makefile arch=linux-ubuntu20.04-haswell [+] ^mpich@4.1.1%clang@16.0.0~argobots~cuda~fortran+hwloc+hydra+libxml2+pci~rocm+romio~slurm~two_level_namespace~vci~verbs+wrapperrpath build_system=autotools datatype-engine=auto device=ch4 netmod=ofi pmi=pmi arch=linux-ubuntu20.04-haswell [+] ^findutils@4.7.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^hwloc@2.9.1%clang@16.0.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell [+] ^ncurses@6.2.20200212%clang@16.0.0+symlinks+termlib abi=6 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libfabric@1.18.0%clang@16.0.0~debug~kdreg build_system=autotools fabrics=sockets,tcp,udp arch=linux-ubuntu20.04-haswell [+] ^libpciaccess@0.17%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libtool@2.4.6%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^util-macros@1.19.3%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libxml2@2.10.3%clang@16.0.0~python build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libiconv@1.17%clang@16.0.0 build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell [+] ^xz@5.2.4%clang@16.0.0~pic build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell [+] ^yaksa@0.2%gcc@12.2.0~cuda~rocm build_system=autotools arch=linux-ubuntu20.04-haswell [-] ^autoconf@2.69%gcc@12.2.0 build_system=autotools patches=7793209 arch=linux-ubuntu20.04-haswell [-] ^automake@1.16.1%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell [-] ^libtool@2.4.6%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell [-] ^m4@1.4.18%gcc@12.2.0+sigsegv build_system=autotools patches=3877ab5,fc9b616 arch=linux-ubuntu20.04-haswell [-] ^python@3.8.10%gcc@12.2.0+bz2+crypt+ctypes+dbm~debug+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tkinter+uuid+zlib build_system=generic patches=0d98e93,4c24573,f2fd060 arch=linux-ubuntu20.04-haswell ``` ### Error message <details><summary>Error message</summary> <pre> ==> Error: AttributeError: 'Spec' object has no attribute 'mpifc' The 'cgns' package cannot find an attribute while trying to build from sources. This might be due to a change in Spack's package format to support multiple build-systems for a single package. You can fix this by updating the build recipe, and you can also report the issue as a bug. More information at https://spack.readthedocs.io/en/latest/packaging_guide.html#installation-procedure /home/nmmoral/opt/spack/var/spack/repos/builtin/packages/cgns/package.py:93, in cmake_args: 90 [ 91 "-DCMAKE_C_COMPILER=%s" % spec["mpi"].mpicc, 92 "-DCMAKE_CXX_COMPILER=%s" % spec["mpi"].mpicxx, >> 93 "-DCMAKE_Fortran_COMPILER=%s" % spec["mpi"].mpifc, 94 ] 95 ) 96 See build log for details: /tmp/nmmoral/spack-stage/spack-stage-cgns-4.3.0-h42e75qpcnab3ozroegcqbrx2ghlqkat/spack-build-out.txt </pre></details> ### Information on your system ``` * **Spack:** 0.20.0.dev0 (f5c32d57e051a8ea6f04bc86f3df5867ea1ab49e) * **Python:** 3.8.10 * **Platform:** linux-ubuntu20.04-skylake_avx512 * **Concretizer:** clingo ``` ### Additional information Maintainer: @gsjaardema Note the bug here is that the recipe for cgns unconditionally sets the fortran compiler to mpifc even if mpifc is not available in the mpi install. spack-build-out.txt ``` ==> cgns: Executing phase: 'cmake' ``` I can't upload my spack-build-env but please contact me if you need some specific info. I don't think it's too relevant to the bug however. ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
1.0
Installation issue: cgns tries to use mpifc even if fortran is disabled - ### Steps to reproduce the issue ```console $ spack spec -I cgns@4.3.0~fortran ^mpich@4.1.1~fortran Input spec -------------------------------- - cgns@4.3.0~fortran - ^mpich@4.1.1~fortran Concretized -------------------------------- - cgns@4.3.0%clang@16.0.0~base_scope~fortran+hdf5~int64~ipo~legacy~mem_debug+mpi+scoping+shared~static~testing~tools build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell [+] ^cmake@3.25.1%clang@16.0.0~doc+ncurses+ownlibs~qt build_system=generic build_type=Release arch=linux-ubuntu20.04-haswell [+] ^gmake@4.2.1%clang@16.0.0~guile build_system=autotools patches=ca60bd9,fe5b60d arch=linux-ubuntu20.04-haswell [+] ^hdf5@1.12.2%clang@16.0.0~cxx~fortran~hl~ipo~java+mpi+shared~szip~threadsafe+tools api=default build_system=cmake build_type=RelWithDebInfo generator=make arch=linux-ubuntu20.04-haswell [+] ^pkgconf@1.8.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^zlib@1.2.13%clang@16.0.0+optimize+pic+shared build_system=makefile arch=linux-ubuntu20.04-haswell [+] ^mpich@4.1.1%clang@16.0.0~argobots~cuda~fortran+hwloc+hydra+libxml2+pci~rocm+romio~slurm~two_level_namespace~vci~verbs+wrapperrpath build_system=autotools datatype-engine=auto device=ch4 netmod=ofi pmi=pmi arch=linux-ubuntu20.04-haswell [+] ^findutils@4.7.0%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^hwloc@2.9.1%clang@16.0.0~cairo~cuda~gl~libudev+libxml2~netloc~nvml~oneapi-level-zero~opencl+pci~rocm build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell [+] ^ncurses@6.2.20200212%clang@16.0.0+symlinks+termlib abi=6 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libfabric@1.18.0%clang@16.0.0~debug~kdreg build_system=autotools fabrics=sockets,tcp,udp arch=linux-ubuntu20.04-haswell [+] ^libpciaccess@0.17%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libtool@2.4.6%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^util-macros@1.19.3%clang@16.0.0 build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libxml2@2.10.3%clang@16.0.0~python build_system=autotools arch=linux-ubuntu20.04-haswell [+] ^libiconv@1.17%clang@16.0.0 build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell [+] ^xz@5.2.4%clang@16.0.0~pic build_system=autotools libs=shared,static arch=linux-ubuntu20.04-haswell [+] ^yaksa@0.2%gcc@12.2.0~cuda~rocm build_system=autotools arch=linux-ubuntu20.04-haswell [-] ^autoconf@2.69%gcc@12.2.0 build_system=autotools patches=7793209 arch=linux-ubuntu20.04-haswell [-] ^automake@1.16.1%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell [-] ^libtool@2.4.6%gcc@12.2.0 build_system=autotools arch=linux-ubuntu20.04-haswell [-] ^m4@1.4.18%gcc@12.2.0+sigsegv build_system=autotools patches=3877ab5,fc9b616 arch=linux-ubuntu20.04-haswell [-] ^python@3.8.10%gcc@12.2.0+bz2+crypt+ctypes+dbm~debug+libxml2+lzma~nis~optimizations+pic+pyexpat+pythoncmd+readline+shared+sqlite3+ssl~tkinter+uuid+zlib build_system=generic patches=0d98e93,4c24573,f2fd060 arch=linux-ubuntu20.04-haswell ``` ### Error message <details><summary>Error message</summary> <pre> ==> Error: AttributeError: 'Spec' object has no attribute 'mpifc' The 'cgns' package cannot find an attribute while trying to build from sources. This might be due to a change in Spack's package format to support multiple build-systems for a single package. You can fix this by updating the build recipe, and you can also report the issue as a bug. More information at https://spack.readthedocs.io/en/latest/packaging_guide.html#installation-procedure /home/nmmoral/opt/spack/var/spack/repos/builtin/packages/cgns/package.py:93, in cmake_args: 90 [ 91 "-DCMAKE_C_COMPILER=%s" % spec["mpi"].mpicc, 92 "-DCMAKE_CXX_COMPILER=%s" % spec["mpi"].mpicxx, >> 93 "-DCMAKE_Fortran_COMPILER=%s" % spec["mpi"].mpifc, 94 ] 95 ) 96 See build log for details: /tmp/nmmoral/spack-stage/spack-stage-cgns-4.3.0-h42e75qpcnab3ozroegcqbrx2ghlqkat/spack-build-out.txt </pre></details> ### Information on your system ``` * **Spack:** 0.20.0.dev0 (f5c32d57e051a8ea6f04bc86f3df5867ea1ab49e) * **Python:** 3.8.10 * **Platform:** linux-ubuntu20.04-skylake_avx512 * **Concretizer:** clingo ``` ### Additional information Maintainer: @gsjaardema Note the bug here is that the recipe for cgns unconditionally sets the fortran compiler to mpifc even if mpifc is not available in the mpi install. spack-build-out.txt ``` ==> cgns: Executing phase: 'cmake' ``` I can't upload my spack-build-env but please contact me if you need some specific info. I don't think it's too relevant to the bug however. ### General information - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [X] I have run `spack maintainers <name-of-the-package>` and **@mentioned** any maintainers - [X] I have uploaded the build log and environment files - [X] I have searched the issues of this repo and believe this is not a duplicate
non_process
installation issue cgns tries to use mpifc even if fortran is disabled steps to reproduce the issue console spack spec i cgns fortran mpich fortran input spec cgns fortran mpich fortran concretized cgns clang base scope fortran ipo legacy mem debug mpi scoping shared static testing tools build system cmake build type relwithdebinfo generator make arch linux haswell cmake clang doc ncurses ownlibs qt build system generic build type release arch linux haswell gmake clang guile build system autotools patches arch linux haswell clang cxx fortran hl ipo java mpi shared szip threadsafe tools api default build system cmake build type relwithdebinfo generator make arch linux haswell pkgconf clang build system autotools arch linux haswell zlib clang optimize pic shared build system makefile arch linux haswell mpich clang argobots cuda fortran hwloc hydra pci rocm romio slurm two level namespace vci verbs wrapperrpath build system autotools datatype engine auto device netmod ofi pmi pmi arch linux haswell findutils clang build system autotools arch linux haswell hwloc clang cairo cuda gl libudev netloc nvml oneapi level zero opencl pci rocm build system autotools libs shared static arch linux haswell ncurses clang symlinks termlib abi build system autotools arch linux haswell libfabric clang debug kdreg build system autotools fabrics sockets tcp udp arch linux haswell libpciaccess clang build system autotools arch linux haswell libtool clang build system autotools arch linux haswell util macros clang build system autotools arch linux haswell clang python build system autotools arch linux haswell libiconv clang build system autotools libs shared static arch linux haswell xz clang pic build system autotools libs shared static arch linux haswell yaksa gcc cuda rocm build system autotools arch linux haswell autoconf gcc build system autotools patches arch linux haswell automake gcc build system autotools arch linux haswell libtool gcc build system autotools arch linux haswell gcc sigsegv build system autotools patches arch linux haswell python gcc crypt ctypes dbm debug lzma nis optimizations pic pyexpat pythoncmd readline shared ssl tkinter uuid zlib build system generic patches arch linux haswell error message error message error attributeerror spec object has no attribute mpifc the cgns package cannot find an attribute while trying to build from sources this might be due to a change in spack s package format to support multiple build systems for a single package you can fix this by updating the build recipe and you can also report the issue as a bug more information at home nmmoral opt spack var spack repos builtin packages cgns package py in cmake args dcmake c compiler s spec mpicc dcmake cxx compiler s spec mpicxx dcmake fortran compiler s spec mpifc see build log for details tmp nmmoral spack stage spack stage cgns spack build out txt information on your system spack python platform linux skylake concretizer clingo additional information maintainer gsjaardema note the bug here is that the recipe for cgns unconditionally sets the fortran compiler to mpifc even if mpifc is not available in the mpi install spack build out txt cgns executing phase cmake i can t upload my spack build env but please contact me if you need some specific info i don t think it s too relevant to the bug however general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
0
1,940
4,769,390,751
IssuesEvent
2016-10-26 12:26:27
openvstorage/alba
https://api.github.com/repos/openvstorage/alba
closed
Alba proxy crashes with Unix_error "Invalid argument" select
priority_critical process_wontfix type_bug
``` alba: internal error, uncaught exception: (Unix.Unix_error "Invalid argument" select "") Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50 Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39 Called from file "src/unix/lwt_main.ml", line 41, characters 8-82 Uncaught exception: (Unix.Unix_error "Invalid argument" select "") Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50 Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39 Called from file "src/unix/lwt_main.ml", line 41, characters 8-82 Called from file "pervasives.ml", line 482, characters 30-33 Called from file "pervasives.ml", line 487, characters 2-15 Called from file "src/alba.ml", line 828, characters 16-22 ``` Saw this 450 times in a log file of a day. The env (OVH) was not configured to use rdma.
1.0
Alba proxy crashes with Unix_error "Invalid argument" select - ``` alba: internal error, uncaught exception: (Unix.Unix_error "Invalid argument" select "") Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50 Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39 Called from file "src/unix/lwt_main.ml", line 41, characters 8-82 Uncaught exception: (Unix.Unix_error "Invalid argument" select "") Raised by primitive operation at file "lwt_rsocket.ml", line 452, characters 12-50 Called from file "src/unix/lwt_engine.ml", line 312, characters 8-39 Called from file "src/unix/lwt_main.ml", line 41, characters 8-82 Called from file "pervasives.ml", line 482, characters 30-33 Called from file "pervasives.ml", line 487, characters 2-15 Called from file "src/alba.ml", line 828, characters 16-22 ``` Saw this 450 times in a log file of a day. The env (OVH) was not configured to use rdma.
process
alba proxy crashes with unix error invalid argument select alba internal error uncaught exception unix unix error invalid argument select raised by primitive operation at file lwt rsocket ml line characters called from file src unix lwt engine ml line characters called from file src unix lwt main ml line characters uncaught exception unix unix error invalid argument select raised by primitive operation at file lwt rsocket ml line characters called from file src unix lwt engine ml line characters called from file src unix lwt main ml line characters called from file pervasives ml line characters called from file pervasives ml line characters called from file src alba ml line characters saw this times in a log file of a day the env ovh was not configured to use rdma
1
18,899
6,655,669,744
IssuesEvent
2017-09-29 17:22:42
habitat-sh/habitat
https://api.github.com/repos/habitat-sh/habitat
opened
Members tab should only be visible to origin members
A-builder C-bug L-javascript
E.g.: https://bldr.habitat.sh/#/origins/cnunciato/members The invitation API calls do fail, of course, but the UI should make it clear you really shouldn't be here.
1.0
Members tab should only be visible to origin members - E.g.: https://bldr.habitat.sh/#/origins/cnunciato/members The invitation API calls do fail, of course, but the UI should make it clear you really shouldn't be here.
non_process
members tab should only be visible to origin members e g the invitation api calls do fail of course but the ui should make it clear you really shouldn t be here
0
16,199
20,710,470,723
IssuesEvent
2022-03-12 00:04:40
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
_handle_zeros_in_scale causing improper scaling when using StandardScaler()
Bug module:preprocessing
<!-- Before submitting a bug, please make sure the issue hasn't been already addressed by searching through the past issues. --> #### Describe the bug There is no floating point tolerance in function _handle_zeros_in_scale for checking if scale == 0.0. As a result, floating point precision can cause this check to incorrectly fail and not set scale to 1.0. The end result is to potentially have an incorrectly scaled values when using StandardScaler() since the value of scale_ will be near 0 instead of 1, introducing numerical instability. #### Steps/Code to Reproduce ``` from sklearn.preprocessing import StandardScaler import numpy as np data_fails = np.full((1000, 1), 14.62, dtype=float).reshape(-1,1) #array filled with 14.62, causes issue data_works = np.full((1000,1), 100.0 , dtype=float).reshape(-1,1) #array filled with 100.0, works as intended scaler_fails = StandardScaler() scaler_works = StandardScaler() scaled_fails = scaler_fails.fit_transform(data_fails) #Returns array filled with -1.0 scaled_works = scaler_works.fit_transform(data_works) #Returns array fill with 0.0 print('\n Results: \n\n') print(scaled_fails[0][0]) print(scaled_works[0][0]) ``` #### Expected Results Expected both scaled results to be zero vector since both are constant-valued vectors. #### Actual Results Standard scaling subtracts mean and divides by standard deviation when appropriate flags are set as in example above. Variance of constant valued vector is 0 which should be caught and replaced by 1 in function _handle_zeros_in_scale. However, this is not happening due variations introduced by floating point representation. Results in mean_ being divided by small floating point value resulting in incorrect scaling when using StandardScaler(). Error occurs at line number 77 in my version of _data inside function _handle_zeros_in_scale. Currently reads: scale[scale == 0.0] = 1.0 #### Versions Python dependencies: pip: 20.0.2 setuptools: 47.1.1.post20200604 sklearn: 0.22.1 numpy: 1.18.1 scipy: 1.4.1 Cython: None pandas: 1.0.3 matplotlib: 3.2.1 joblib: 0.15.1 <!-- Thanks for contributing! -->
1.0
_handle_zeros_in_scale causing improper scaling when using StandardScaler() - <!-- Before submitting a bug, please make sure the issue hasn't been already addressed by searching through the past issues. --> #### Describe the bug There is no floating point tolerance in function _handle_zeros_in_scale for checking if scale == 0.0. As a result, floating point precision can cause this check to incorrectly fail and not set scale to 1.0. The end result is to potentially have an incorrectly scaled values when using StandardScaler() since the value of scale_ will be near 0 instead of 1, introducing numerical instability. #### Steps/Code to Reproduce ``` from sklearn.preprocessing import StandardScaler import numpy as np data_fails = np.full((1000, 1), 14.62, dtype=float).reshape(-1,1) #array filled with 14.62, causes issue data_works = np.full((1000,1), 100.0 , dtype=float).reshape(-1,1) #array filled with 100.0, works as intended scaler_fails = StandardScaler() scaler_works = StandardScaler() scaled_fails = scaler_fails.fit_transform(data_fails) #Returns array filled with -1.0 scaled_works = scaler_works.fit_transform(data_works) #Returns array fill with 0.0 print('\n Results: \n\n') print(scaled_fails[0][0]) print(scaled_works[0][0]) ``` #### Expected Results Expected both scaled results to be zero vector since both are constant-valued vectors. #### Actual Results Standard scaling subtracts mean and divides by standard deviation when appropriate flags are set as in example above. Variance of constant valued vector is 0 which should be caught and replaced by 1 in function _handle_zeros_in_scale. However, this is not happening due variations introduced by floating point representation. Results in mean_ being divided by small floating point value resulting in incorrect scaling when using StandardScaler(). Error occurs at line number 77 in my version of _data inside function _handle_zeros_in_scale. Currently reads: scale[scale == 0.0] = 1.0 #### Versions Python dependencies: pip: 20.0.2 setuptools: 47.1.1.post20200604 sklearn: 0.22.1 numpy: 1.18.1 scipy: 1.4.1 Cython: None pandas: 1.0.3 matplotlib: 3.2.1 joblib: 0.15.1 <!-- Thanks for contributing! -->
process
handle zeros in scale causing improper scaling when using standardscaler before submitting a bug please make sure the issue hasn t been already addressed by searching through the past issues describe the bug there is no floating point tolerance in function handle zeros in scale for checking if scale as a result floating point precision can cause this check to incorrectly fail and not set scale to the end result is to potentially have an incorrectly scaled values when using standardscaler since the value of scale will be near instead of introducing numerical instability steps code to reproduce from sklearn preprocessing import standardscaler import numpy as np data fails np full dtype float reshape array filled with causes issue data works np full dtype float reshape array filled with works as intended scaler fails standardscaler scaler works standardscaler scaled fails scaler fails fit transform data fails returns array filled with scaled works scaler works fit transform data works returns array fill with print n results n n print scaled fails print scaled works expected results expected both scaled results to be zero vector since both are constant valued vectors actual results standard scaling subtracts mean and divides by standard deviation when appropriate flags are set as in example above variance of constant valued vector is which should be caught and replaced by in function handle zeros in scale however this is not happening due variations introduced by floating point representation results in mean being divided by small floating point value resulting in incorrect scaling when using standardscaler error occurs at line number in my version of data inside function handle zeros in scale currently reads scale versions python dependencies pip setuptools sklearn numpy scipy cython none pandas matplotlib joblib
1
12,423
14,923,397,384
IssuesEvent
2021-01-23 18:46:06
AcademySoftwareFoundation/OpenCue
https://api.github.com/repos/AcademySoftwareFoundation/OpenCue
opened
docker files need to pin max pip version for python2
process
**Describe the process** The latest version of `pip` (21.0) and `setuptools (45) has just dropped python 2 support. As a result building the sandbox environment fails: ``` docker-compose --project-directory . -f sandbox/docker-compose.yml build ``` ``` Step 8/25 : RUN python -m pip install --upgrade setuptools ---> Running in 8eb1abc898a2 Traceback (most recent call last): File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/usr/lib/python2.7/site-packages/pip/__main__.py", line 21, in <module> from pip._internal.cli.main import main as _main File "/usr/lib/python2.7/site-packages/pip/_internal/cli/main.py", line 60 sys.stderr.write(f"ERROR: {exc}") ^ SyntaxError: invalid syntax ERROR: Service 'metrics' failed to build : The command '/bin/sh -c python -m pip install --upgrade setuptools' returned a non-zero code: 1 ``` We need should pin `pip` < 21 and `setuptools` < 45 during the calls to upgrade these packages in the various Dockerfiles ``` RUN python -m pip install --upgrade 'pip<21' RUN python -m pip install --upgrade 'setuptools<45' ```
1.0
docker files need to pin max pip version for python2 - **Describe the process** The latest version of `pip` (21.0) and `setuptools (45) has just dropped python 2 support. As a result building the sandbox environment fails: ``` docker-compose --project-directory . -f sandbox/docker-compose.yml build ``` ``` Step 8/25 : RUN python -m pip install --upgrade setuptools ---> Running in 8eb1abc898a2 Traceback (most recent call last): File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/usr/lib/python2.7/site-packages/pip/__main__.py", line 21, in <module> from pip._internal.cli.main import main as _main File "/usr/lib/python2.7/site-packages/pip/_internal/cli/main.py", line 60 sys.stderr.write(f"ERROR: {exc}") ^ SyntaxError: invalid syntax ERROR: Service 'metrics' failed to build : The command '/bin/sh -c python -m pip install --upgrade setuptools' returned a non-zero code: 1 ``` We need should pin `pip` < 21 and `setuptools` < 45 during the calls to upgrade these packages in the various Dockerfiles ``` RUN python -m pip install --upgrade 'pip<21' RUN python -m pip install --upgrade 'setuptools<45' ```
process
docker files need to pin max pip version for describe the process the latest version of pip and setuptools has just dropped python support as a result building the sandbox environment fails docker compose project directory f sandbox docker compose yml build step run python m pip install upgrade setuptools running in traceback most recent call last file usr runpy py line in run module as main main fname loader pkg name file usr runpy py line in run code exec code in run globals file usr lib site packages pip main py line in from pip internal cli main import main as main file usr lib site packages pip internal cli main py line sys stderr write f error exc syntaxerror invalid syntax error service metrics failed to build the command bin sh c python m pip install upgrade setuptools returned a non zero code we need should pin pip and setuptools during the calls to upgrade these packages in the various dockerfiles run python m pip install upgrade pip run python m pip install upgrade setuptools
1
22,713
32,038,297,439
IssuesEvent
2023-09-22 17:05:17
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
nx 16.8.1 has 2 guarddog issues
npm-install-script npm-silent-process-execution
```{"npm-install-script":[{"code":" \"postinstall\": \"node ./bin/post-install\"","location":"package/package.json:12","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const p = (0, child_process_1.spawn)('node', [scriptPath, `\"${this.cachePath}\"`], {\n stdio: 'ignore',\n detached: true,\n shell: false,\n });","location":"package/src/tasks-runner/cache.js:28","message":"This package is silently executing another executable"}]}```
1.0
nx 16.8.1 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"postinstall\": \"node ./bin/post-install\"","location":"package/package.json:12","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const p = (0, child_process_1.spawn)('node', [scriptPath, `\"${this.cachePath}\"`], {\n stdio: 'ignore',\n detached: true,\n shell: false,\n });","location":"package/src/tasks-runner/cache.js:28","message":"This package is silently executing another executable"}]}```
process
nx has guarddog issues npm install script npm silent process execution n stdio ignore n detached true n shell false n location package src tasks runner cache js message this package is silently executing another executable
1
17,984
24,006,190,811
IssuesEvent
2022-09-14 14:58:22
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/k8sattributes] agent passthrough, gateway k8s.pod.ip configuration broken post v0.55.0 release
bug priority:p2 processor/k8sattributes
**Describe the bug** In an agent/gateway k8s deployment, the `k8sattributes` processor no longer correctly adds k8s metadata to the Resource associated with traces emitted by a pod elsewhere in the cluster. I suspect this may have broken due to [PR #8465](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/8465) **Steps to reproduce** Create a daemonset agent/deployment gateway collector setup in a k8s cluster with the configuration files shown below. Generate traces from a pod that sends to the local daemonset agent pod. Observe the debug logs on the gateway collector. **What did you expect to see?** Logs for each ResourceSpans indicating each Resource with a preexisting `k8s.pod.ip` attribute had the Resource augmented with `k8s.namespace.name`, `k8s.pod.uid` and `k8s.pod.name`. **What did you see instead?** Logs showing the incoming traces have a `Resource` with the correct pod IP in the `k8s.pod.ip` attribute, but **none** of `k8s.namespace.name`, `k8s.pod.uid` nor `k8s.pod.name`. **What version did you use?** Version: v0.56.0 **What config did you use?** On the agent side: ```yaml receivers: otlp: protocols: grpc: 0.0.0.0:4317 processors: k8sattributes: passthrough: true batch: exporters: otlp: endpoint: otel-gateway:4317 service: pipelines: traces: receivers: [ otlp ] processors: [ k8sattributes, batch ] exporters: [ otlp ] ``` and on the gateway side ```yaml receivers: otlp: protocols: grpc: 0.0.0.0:4317 processors: batch: k8sattributes: pod_association: - sources: - from: resource_attribute name: k8s.pod.ip extract: metadata: - k8s.namespace.name - k8s.pod.name - k8s.pod.uid exporters: logging: loglevel: debug service: pipelines: traces: receivers: [ otlp ] processors: [ k8sattributes, batch ] exporters: [ logging ] ``` **Environment** OS: RHEL8 Linux Compiler(if manually compiled): go1.17 **Additional context** From the extension debug logs, the incoming Resource is already augmented by the agent pod with `k8s.pod.ip=<podIP>`. In `kubernetesprocessor.processResource()`, the debug message `"evaluating pod identifier [ {Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podIp>" }, ..]"` is printed, but it appears to return false from `WatchClient.GetPod(<podIdentifier>)`. This seems to indicate that there is no entry in `c.Pods[]` for that specific PodIdentifier. The primary place in the code where pods are added to the `c.Pods[]` map is in `WatchClient.addOrUpdatePod()`. Prior to the aforementioned PR, the keys for the Pods map were Strings containing either the IP address or the pod UID, and there was no distinction between whether it came from the "connection" or from a "resource_attribute". After the PR, the key(s) returned from`WatchClient.getIdentifiersFromAssoc()` are inserted into the Pods map, with each key pointing to the pod info. To correctly match the key which is printed in `evaluating pod identifier...`, `getIdentifiersFromAssoc()` should return the key `{Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podip>" }` when given the input association specification, namely: ```yaml pod_association: - sources: - from: resource_attribute name: k8s.pod.ip ``` Based on code inspection, it appears this does not happen. I suspect changing the [switch statement in `getIdentifiersFromAssoc()`](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/fffc27fe3a5f360289d21604e61779899ef0fefa/processor/k8sattributesprocessor/internal/kube/client.go#L468-L483) from: ```go case conventions.AttributeHostName: attr = pod.Address ``` to ```go case conventions.AttributeHostName, "k8s.pod.ip": attr = pod.Address ``` would probably fix this problem. But I have only ascertained this via code inspection, not compiling and running the code! Possibly related: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/13119
1.0
[processor/k8sattributes] agent passthrough, gateway k8s.pod.ip configuration broken post v0.55.0 release - **Describe the bug** In an agent/gateway k8s deployment, the `k8sattributes` processor no longer correctly adds k8s metadata to the Resource associated with traces emitted by a pod elsewhere in the cluster. I suspect this may have broken due to [PR #8465](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/8465) **Steps to reproduce** Create a daemonset agent/deployment gateway collector setup in a k8s cluster with the configuration files shown below. Generate traces from a pod that sends to the local daemonset agent pod. Observe the debug logs on the gateway collector. **What did you expect to see?** Logs for each ResourceSpans indicating each Resource with a preexisting `k8s.pod.ip` attribute had the Resource augmented with `k8s.namespace.name`, `k8s.pod.uid` and `k8s.pod.name`. **What did you see instead?** Logs showing the incoming traces have a `Resource` with the correct pod IP in the `k8s.pod.ip` attribute, but **none** of `k8s.namespace.name`, `k8s.pod.uid` nor `k8s.pod.name`. **What version did you use?** Version: v0.56.0 **What config did you use?** On the agent side: ```yaml receivers: otlp: protocols: grpc: 0.0.0.0:4317 processors: k8sattributes: passthrough: true batch: exporters: otlp: endpoint: otel-gateway:4317 service: pipelines: traces: receivers: [ otlp ] processors: [ k8sattributes, batch ] exporters: [ otlp ] ``` and on the gateway side ```yaml receivers: otlp: protocols: grpc: 0.0.0.0:4317 processors: batch: k8sattributes: pod_association: - sources: - from: resource_attribute name: k8s.pod.ip extract: metadata: - k8s.namespace.name - k8s.pod.name - k8s.pod.uid exporters: logging: loglevel: debug service: pipelines: traces: receivers: [ otlp ] processors: [ k8sattributes, batch ] exporters: [ logging ] ``` **Environment** OS: RHEL8 Linux Compiler(if manually compiled): go1.17 **Additional context** From the extension debug logs, the incoming Resource is already augmented by the agent pod with `k8s.pod.ip=<podIP>`. In `kubernetesprocessor.processResource()`, the debug message `"evaluating pod identifier [ {Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podIp>" }, ..]"` is printed, but it appears to return false from `WatchClient.GetPod(<podIdentifier>)`. This seems to indicate that there is no entry in `c.Pods[]` for that specific PodIdentifier. The primary place in the code where pods are added to the `c.Pods[]` map is in `WatchClient.addOrUpdatePod()`. Prior to the aforementioned PR, the keys for the Pods map were Strings containing either the IP address or the pod UID, and there was no distinction between whether it came from the "connection" or from a "resource_attribute". After the PR, the key(s) returned from`WatchClient.getIdentifiersFromAssoc()` are inserted into the Pods map, with each key pointing to the pod info. To correctly match the key which is printed in `evaluating pod identifier...`, `getIdentifiersFromAssoc()` should return the key `{Source: { From: "resource_attribute", Name: "k8s.pod.ip" }, Value: "<podip>" }` when given the input association specification, namely: ```yaml pod_association: - sources: - from: resource_attribute name: k8s.pod.ip ``` Based on code inspection, it appears this does not happen. I suspect changing the [switch statement in `getIdentifiersFromAssoc()`](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/fffc27fe3a5f360289d21604e61779899ef0fefa/processor/k8sattributesprocessor/internal/kube/client.go#L468-L483) from: ```go case conventions.AttributeHostName: attr = pod.Address ``` to ```go case conventions.AttributeHostName, "k8s.pod.ip": attr = pod.Address ``` would probably fix this problem. But I have only ascertained this via code inspection, not compiling and running the code! Possibly related: https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/13119
process
agent passthrough gateway pod ip configuration broken post release describe the bug in an agent gateway deployment the processor no longer correctly adds metadata to the resource associated with traces emitted by a pod elsewhere in the cluster i suspect this may have broken due to steps to reproduce create a daemonset agent deployment gateway collector setup in a cluster with the configuration files shown below generate traces from a pod that sends to the local daemonset agent pod observe the debug logs on the gateway collector what did you expect to see logs for each resourcespans indicating each resource with a preexisting pod ip attribute had the resource augmented with namespace name pod uid and pod name what did you see instead logs showing the incoming traces have a resource with the correct pod ip in the pod ip attribute but none of namespace name pod uid nor pod name what version did you use version what config did you use on the agent side yaml receivers otlp protocols grpc processors passthrough true batch exporters otlp endpoint otel gateway service pipelines traces receivers processors exporters and on the gateway side yaml receivers otlp protocols grpc processors batch pod association sources from resource attribute name pod ip extract metadata namespace name pod name pod uid exporters logging loglevel debug service pipelines traces receivers processors exporters environment os linux compiler if manually compiled additional context from the extension debug logs the incoming resource is already augmented by the agent pod with pod ip in kubernetesprocessor processresource the debug message evaluating pod identifier is printed but it appears to return false from watchclient getpod this seems to indicate that there is no entry in c pods for that specific podidentifier the primary place in the code where pods are added to the c pods map is in watchclient addorupdatepod prior to the aforementioned pr the keys for the pods map were strings containing either the ip address or the pod uid and there was no distinction between whether it came from the connection or from a resource attribute after the pr the key s returned from watchclient getidentifiersfromassoc are inserted into the pods map with each key pointing to the pod info to correctly match the key which is printed in evaluating pod identifier getidentifiersfromassoc should return the key source from resource attribute name pod ip value when given the input association specification namely yaml pod association sources from resource attribute name pod ip based on code inspection it appears this does not happen i suspect changing the from go case conventions attributehostname attr pod address to go case conventions attributehostname pod ip attr pod address would probably fix this problem but i have only ascertained this via code inspection not compiling and running the code possibly related
1
15,149
18,906,310,222
IssuesEvent
2021-11-16 09:27:42
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
NTR: Negative Regulation of CRISPR-cas system
multi-species process
New Term: Negative regulation of CRISPR-cas system Definition: Negative regulation of an adaptive immune response of bacteria that serves to clear host cells of foreign DNA and RNA. It has three distinct stage: acquisition of foreign DNA by integration into CRISPR loci in the host chromosome, CRISPR RNA (crRNA) biogenesis, and target interference. CISPR stands for Clustered Regularly Interspaced Short Palindromic Repeat, which describes the nature of the loci. Notes: This reference describes the function of anti-CRISPR proteins within a bacterial genome to protect itself from its CRISPR system. PMID: 30190308, 30190307 @sandyl27 @jimhu-tamu This is for CACAO
1.0
NTR: Negative Regulation of CRISPR-cas system - New Term: Negative regulation of CRISPR-cas system Definition: Negative regulation of an adaptive immune response of bacteria that serves to clear host cells of foreign DNA and RNA. It has three distinct stage: acquisition of foreign DNA by integration into CRISPR loci in the host chromosome, CRISPR RNA (crRNA) biogenesis, and target interference. CISPR stands for Clustered Regularly Interspaced Short Palindromic Repeat, which describes the nature of the loci. Notes: This reference describes the function of anti-CRISPR proteins within a bacterial genome to protect itself from its CRISPR system. PMID: 30190308, 30190307 @sandyl27 @jimhu-tamu This is for CACAO
process
ntr negative regulation of crispr cas system new term negative regulation of crispr cas system definition negative regulation of an adaptive immune response of bacteria that serves to clear host cells of foreign dna and rna it has three distinct stage acquisition of foreign dna by integration into crispr loci in the host chromosome crispr rna crrna biogenesis and target interference cispr stands for clustered regularly interspaced short palindromic repeat which describes the nature of the loci notes this reference describes the function of anti crispr proteins within a bacterial genome to protect itself from its crispr system pmid jimhu tamu this is for cacao
1
10,684
13,464,983,681
IssuesEvent
2020-09-09 20:05:40
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
Vagrant Cloud post-processor upload errors with unexpected EOF
bug post-processor/vagrant-cloud
#### Overview of the Issue tl;dr: Vagrant Cloud post-processor fails with an unexpected EOF when uploading a successful build targeting Virtual Box. The build is running on a Jenkins agent. I'm running the build locally at the moment, and will report back with results when it finishes. I'm using a near-minimal Packer template with the Vagrant builder to run an Ansible build on the `bento/ubuntu-18.04` base box. Template is attached below. I've stripped it down to the bare essentials and reduced the Ansible provisioning to a series of package installs with `apt` to attempt to pinpoint the error, so there aren't many distractions. #### Reproduction Steps Trigger the build on a Jenkins agent via Blue Ocean pipelines. See infrastructure details below for details. ### Packer version Packer version `1.4.4`. ### Simplified Packer Buildfile <details> <summary><b>Packer template</b> (`UbuntuVM.json`)</summary> <pre> { "variables": { "cloud_token": "{{ env `VAGRANT_CLOUD_TOKEN` }}", "version": "1.0.{{timestamp}}" }, "builders": [{ "type": "vagrant", "communicator": "ssh", "provider": "virtualbox", "source_path": "bento/ubuntu-18.04", "add_force": true }], "provisioners": [{ "type": "ansible", "playbook_file": "../provisioners/main.yml" }], "post-processors": [ { "type": "vagrant-cloud", "box_tag": "cybersecurity/UbuntuVM", "access_token": "{{user `VAGRANT_CLOUD_TOKEN`}}", "version": "{{user `version`}}" } ] } </pre> </details> <details> <summary><b>Jenkinsfile</b></summary> <pre> pipeline { agent any stages { stage('Build') { steps { sh(label: 'Check directory', script: 'pwd && whoami') sh(label: 'Verify Packer is on $PATH', script: '/usr/local/bin/packer --version') } } stage('Packer') { steps { echo 'Packing...' sh(label: 'Packer Build Stage!', script: '/usr/local/bin/packer build -debug UbuntuVM/packer/UbuntuVM.json') } } stage('Deploy') { steps { echo 'Deploying...' } } } environment { VAGRANT_CLOUD_TOKEN = 'Omitted Cause...Secrets' } } </pre> </details> ### Operating system and Environment details **Infrastructure** - Jenkins v2.203 running on Ubuntu 18.04, deployed to an Azure `Standard_D4S_v3` instance - The Ubuntu instance is unmodified _except_ for the following installations: - Vagrant v2.0.2 - Ansible v2.9.0 - Packer v1.4.4 **Build Details** - The build is described in a declarative `Jenkinsfile` — see above. - Fwiw, webhooks are turned off at the moment, and the build is triggered manually. Builds, for now, are run directly on the Jenkins master as opposed to on distributed agents or within Docker containers. ### Log Fragments and crash.log files See the [gist](https://gist.github.com/Peleke/daea9cb95c4003f6440a35feb095b9c5) for full log output. ### Additional Details / Some Questions The error message I get is: ``` Error uploading box! Will retry in 10 seconds. Error: Put https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ: EOF ``` ...Which is thrown from [packer/post-processor/vagrant-cloud/step_upload.go](https://github.com/hashicorp/packer/blob/dcd21d37da053ccb2fc194c77e710ab6eb5967bf/post-processor/vagrant-cloud/step_upload.go#L17). Unfortunately, I get no additional details regarding the nature of the error: ``` 2019/11/05 16:57:15 packer: 2019/11/05 16:57:15 Post-Processor Vagrant Cloud API Upload: output-vagrant/package.box https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ 2019/11/05 16:58:25 packer: 2019/11/05 16:58:25 Post-Processor Vagrant Cloud Upload Response: 2019/11/05 16:58:25 packer: 2019/11/05 16:58:25 packer: <nil> ``` My best guess (and forgive me if it's either obvious or a boneheaded hunch) is that Vagrant Cloud expects an upload of a given size but reads an EOF prior to receiving that many bytes. I noticed these seemingly related, as-yet unresolved issues: - [#3378](https://github.com/hashicorp/packer/issues/3378) - #8229 If this is just my mistake, please let me know what I'm doing wrong. But if it's actually a bug and someone can point me in the right direction as to the nature of the error, I'd be happy to discuss contributing a fix. Thanks for the time/attention :boom:
1.0
Vagrant Cloud post-processor upload errors with unexpected EOF - #### Overview of the Issue tl;dr: Vagrant Cloud post-processor fails with an unexpected EOF when uploading a successful build targeting Virtual Box. The build is running on a Jenkins agent. I'm running the build locally at the moment, and will report back with results when it finishes. I'm using a near-minimal Packer template with the Vagrant builder to run an Ansible build on the `bento/ubuntu-18.04` base box. Template is attached below. I've stripped it down to the bare essentials and reduced the Ansible provisioning to a series of package installs with `apt` to attempt to pinpoint the error, so there aren't many distractions. #### Reproduction Steps Trigger the build on a Jenkins agent via Blue Ocean pipelines. See infrastructure details below for details. ### Packer version Packer version `1.4.4`. ### Simplified Packer Buildfile <details> <summary><b>Packer template</b> (`UbuntuVM.json`)</summary> <pre> { "variables": { "cloud_token": "{{ env `VAGRANT_CLOUD_TOKEN` }}", "version": "1.0.{{timestamp}}" }, "builders": [{ "type": "vagrant", "communicator": "ssh", "provider": "virtualbox", "source_path": "bento/ubuntu-18.04", "add_force": true }], "provisioners": [{ "type": "ansible", "playbook_file": "../provisioners/main.yml" }], "post-processors": [ { "type": "vagrant-cloud", "box_tag": "cybersecurity/UbuntuVM", "access_token": "{{user `VAGRANT_CLOUD_TOKEN`}}", "version": "{{user `version`}}" } ] } </pre> </details> <details> <summary><b>Jenkinsfile</b></summary> <pre> pipeline { agent any stages { stage('Build') { steps { sh(label: 'Check directory', script: 'pwd && whoami') sh(label: 'Verify Packer is on $PATH', script: '/usr/local/bin/packer --version') } } stage('Packer') { steps { echo 'Packing...' sh(label: 'Packer Build Stage!', script: '/usr/local/bin/packer build -debug UbuntuVM/packer/UbuntuVM.json') } } stage('Deploy') { steps { echo 'Deploying...' } } } environment { VAGRANT_CLOUD_TOKEN = 'Omitted Cause...Secrets' } } </pre> </details> ### Operating system and Environment details **Infrastructure** - Jenkins v2.203 running on Ubuntu 18.04, deployed to an Azure `Standard_D4S_v3` instance - The Ubuntu instance is unmodified _except_ for the following installations: - Vagrant v2.0.2 - Ansible v2.9.0 - Packer v1.4.4 **Build Details** - The build is described in a declarative `Jenkinsfile` — see above. - Fwiw, webhooks are turned off at the moment, and the build is triggered manually. Builds, for now, are run directly on the Jenkins master as opposed to on distributed agents or within Docker containers. ### Log Fragments and crash.log files See the [gist](https://gist.github.com/Peleke/daea9cb95c4003f6440a35feb095b9c5) for full log output. ### Additional Details / Some Questions The error message I get is: ``` Error uploading box! Will retry in 10 seconds. Error: Put https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ: EOF ``` ...Which is thrown from [packer/post-processor/vagrant-cloud/step_upload.go](https://github.com/hashicorp/packer/blob/dcd21d37da053ccb2fc194c77e710ab6eb5967bf/post-processor/vagrant-cloud/step_upload.go#L17). Unfortunately, I get no additional details regarding the nature of the error: ``` 2019/11/05 16:57:15 packer: 2019/11/05 16:57:15 Post-Processor Vagrant Cloud API Upload: output-vagrant/package.box https://archivist.vagrantup.com/v1/object/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJrZXkiOiJib3hlcy82MTBlYTY1OS1kNWRlLTQ2MjctODczNy0yY2I3MjRlNjIxMzkiLCJtb2RlIjoidyIsImV4cGlyZSI6MTU3Mjk3Mzc3NiwiY2FsbGJhY2siOiJodHRwczovL3ZhZ3JhbnRjbG91ZC5jb20vYXBpL2ludGVybmFsL2FyY2hpdmlzdC9jYWxsYmFjayJ9.s5DFfIzBBMFo-4MzMy4SXYcOZ0v79pZqea5a0RSbfqQ 2019/11/05 16:58:25 packer: 2019/11/05 16:58:25 Post-Processor Vagrant Cloud Upload Response: 2019/11/05 16:58:25 packer: 2019/11/05 16:58:25 packer: <nil> ``` My best guess (and forgive me if it's either obvious or a boneheaded hunch) is that Vagrant Cloud expects an upload of a given size but reads an EOF prior to receiving that many bytes. I noticed these seemingly related, as-yet unresolved issues: - [#3378](https://github.com/hashicorp/packer/issues/3378) - #8229 If this is just my mistake, please let me know what I'm doing wrong. But if it's actually a bug and someone can point me in the right direction as to the nature of the error, I'd be happy to discuss contributing a fix. Thanks for the time/attention :boom:
process
vagrant cloud post processor upload errors with unexpected eof overview of the issue tl dr vagrant cloud post processor fails with an unexpected eof when uploading a successful build targeting virtual box the build is running on a jenkins agent i m running the build locally at the moment and will report back with results when it finishes i m using a near minimal packer template with the vagrant builder to run an ansible build on the bento ubuntu base box template is attached below i ve stripped it down to the bare essentials and reduced the ansible provisioning to a series of package installs with apt to attempt to pinpoint the error so there aren t many distractions reproduction steps trigger the build on a jenkins agent via blue ocean pipelines see infrastructure details below for details packer version packer version simplified packer buildfile packer template ubuntuvm json variables cloud token env vagrant cloud token version timestamp builders type vagrant communicator ssh provider virtualbox source path bento ubuntu add force true provisioners type ansible playbook file provisioners main yml post processors type vagrant cloud box tag cybersecurity ubuntuvm access token user vagrant cloud token version user version jenkinsfile pipeline agent any stages stage build steps sh label check directory script pwd whoami sh label verify packer is on path script usr local bin packer version stage packer steps echo packing sh label packer build stage script usr local bin packer build debug ubuntuvm packer ubuntuvm json stage deploy steps echo deploying environment vagrant cloud token omitted cause secrets operating system and environment details infrastructure jenkins running on ubuntu deployed to an azure standard instance the ubuntu instance is unmodified except for the following installations vagrant ansible packer build details the build is described in a declarative jenkinsfile — see above fwiw webhooks are turned off at the moment and the build is triggered manually builds for now are run directly on the jenkins master as opposed to on distributed agents or within docker containers log fragments and crash log files see the for full log output additional details some questions the error message i get is error uploading box will retry in seconds error put eof which is thrown from unfortunately i get no additional details regarding the nature of the error packer post processor vagrant cloud api upload output vagrant package box packer post processor vagrant cloud upload response packer packer my best guess and forgive me if it s either obvious or a boneheaded hunch is that vagrant cloud expects an upload of a given size but reads an eof prior to receiving that many bytes i noticed these seemingly related as yet unresolved issues if this is just my mistake please let me know what i m doing wrong but if it s actually a bug and someone can point me in the right direction as to the nature of the error i d be happy to discuss contributing a fix thanks for the time attention boom
1
5,950
8,775,089,000
IssuesEvent
2018-12-18 21:52:49
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Static Analysis & Code Metrics
difficulty-04-quackhead enhancement parse-tree-processing up-for-grabs user-interface
Would be cool to support metrics and graphs over time. Can start with "simple" metrics like Lines of Code, Cyclomatic Complexity, Maintainability Index, number of broken rules, etc.
1.0
Static Analysis & Code Metrics - Would be cool to support metrics and graphs over time. Can start with "simple" metrics like Lines of Code, Cyclomatic Complexity, Maintainability Index, number of broken rules, etc.
process
static analysis code metrics would be cool to support metrics and graphs over time can start with simple metrics like lines of code cyclomatic complexity maintainability index number of broken rules etc
1
20,816
27,578,247,130
IssuesEvent
2023-03-08 14:31:53
ukri-excalibur/excalibur-tests
https://api.github.com/repos/ukri-excalibur/excalibur-tests
opened
Create high-level script to run postprocessing
UCL postprocessing
This should accept some input either in command line or a config yaml/json file, and based on that it should perform the required type of analysis (see use cases in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139). The steps of this analysis should be something like - read desired data from some file path and store in a pandas dataframe - produce required output (plots and/or tables)
1.0
Create high-level script to run postprocessing - This should accept some input either in command line or a config yaml/json file, and based on that it should perform the required type of analysis (see use cases in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139). The steps of this analysis should be something like - read desired data from some file path and store in a pandas dataframe - produce required output (plots and/or tables)
process
create high level script to run postprocessing this should accept some input either in command line or a config yaml json file and based on that it should perform the required type of analysis see use cases in the steps of this analysis should be something like read desired data from some file path and store in a pandas dataframe produce required output plots and or tables
1
123,713
12,216,324,599
IssuesEvent
2020-05-01 14:54:57
monoDriveIO/documentation
https://api.github.com/repos/monoDriveIO/documentation
opened
Create generic API document
documentation enhancement
Create a generic API introduction that discusses how to send/receive messages to the simulator. This should focus on emphasizing that only messaging is required to actually control the simulation in any language.
1.0
Create generic API document - Create a generic API introduction that discusses how to send/receive messages to the simulator. This should focus on emphasizing that only messaging is required to actually control the simulation in any language.
non_process
create generic api document create a generic api introduction that discusses how to send receive messages to the simulator this should focus on emphasizing that only messaging is required to actually control the simulation in any language
0
14,553
17,670,747,059
IssuesEvent
2021-08-23 05:35:13
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[feature] add Fill noData cells algorithm and tests
Automatic new feature Processing Alg 3.14
Original commit: https://github.com/qgis/QGIS/commit/ea7167e135b47174b240ae4306406a7b3ee8189f by nyalldawson Unfortunately this naughty coder did not write a description... :-(
1.0
[feature] add Fill noData cells algorithm and tests - Original commit: https://github.com/qgis/QGIS/commit/ea7167e135b47174b240ae4306406a7b3ee8189f by nyalldawson Unfortunately this naughty coder did not write a description... :-(
process
add fill nodata cells algorithm and tests original commit by nyalldawson unfortunately this naughty coder did not write a description
1
250,554
7,978,703,077
IssuesEvent
2018-07-17 19:11:32
lupino3/edumips64
https://api.github.com/repos/lupino3/edumips64
opened
Get rid of JavaHelp in favor of JavaFX WebView
priority:1 type:enhancement
Since Java 8, the JRE includes a [WebView component](https://docs.oracle.com/javase/8/javafx/api/javafx/scene/web/WebView.html), which should be easy to integrate in a standard Swing frame (see [JavaFX for Swing](https://docs.oracle.com/javafx/2/swing/swing-fx-interoperability.htm)). Since EduMIPS64 already depends on Java 8+, it would be great to get rid of JavaHelp, which is a dying technology born 10+ years ago and which is as of today the only external dependency for the simulator.
1.0
Get rid of JavaHelp in favor of JavaFX WebView - Since Java 8, the JRE includes a [WebView component](https://docs.oracle.com/javase/8/javafx/api/javafx/scene/web/WebView.html), which should be easy to integrate in a standard Swing frame (see [JavaFX for Swing](https://docs.oracle.com/javafx/2/swing/swing-fx-interoperability.htm)). Since EduMIPS64 already depends on Java 8+, it would be great to get rid of JavaHelp, which is a dying technology born 10+ years ago and which is as of today the only external dependency for the simulator.
non_process
get rid of javahelp in favor of javafx webview since java the jre includes a which should be easy to integrate in a standard swing frame see since already depends on java it would be great to get rid of javahelp which is a dying technology born years ago and which is as of today the only external dependency for the simulator
0
167,807
20,726,409,066
IssuesEvent
2022-03-14 02:49:30
dgee2/dgee2.github.io
https://api.github.com/repos/dgee2/dgee2.github.io
opened
CVE-2021-33502 (High) detected in normalize-url-4.5.0.tgz
security vulnerability
## CVE-2021-33502 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-4.5.0.tgz</b></p></summary> <p>Normalize a URL</p> <p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/gatsby-plugin-sharp/node_modules/normalize-url/package.json,/node_modules/package-json/node_modules/normalize-url/package.json,/node_modules/gatsby-source-filesystem/node_modules/normalize-url/package.json</p> <p> Dependency Hierarchy: - gatsby-plugin-sharp-3.2.0.tgz (Root Library) - got-10.7.0.tgz - cacheable-request-7.0.1.tgz - :x: **normalize-url-4.5.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs. <p>Publish Date: 2021-05-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p> <p>Release Date: 2021-05-24</p> <p>Fix Resolution (normalize-url): 4.5.1</p> <p>Direct dependency fix Resolution (gatsby-plugin-sharp): 3.2.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-33502 (High) detected in normalize-url-4.5.0.tgz - ## CVE-2021-33502 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-4.5.0.tgz</b></p></summary> <p>Normalize a URL</p> <p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-4.5.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/gatsby-plugin-sharp/node_modules/normalize-url/package.json,/node_modules/package-json/node_modules/normalize-url/package.json,/node_modules/gatsby-source-filesystem/node_modules/normalize-url/package.json</p> <p> Dependency Hierarchy: - gatsby-plugin-sharp-3.2.0.tgz (Root Library) - got-10.7.0.tgz - cacheable-request-7.0.1.tgz - :x: **normalize-url-4.5.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs. <p>Publish Date: 2021-05-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p> <p>Release Date: 2021-05-24</p> <p>Fix Resolution (normalize-url): 4.5.1</p> <p>Direct dependency fix Resolution (gatsby-plugin-sharp): 3.2.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in normalize url tgz cve high severity vulnerability vulnerable library normalize url tgz normalize a url library home page a href path to dependency file package json path to vulnerable library node modules gatsby plugin sharp node modules normalize url package json node modules package json node modules normalize url package json node modules gatsby source filesystem node modules normalize url package json dependency hierarchy gatsby plugin sharp tgz root library got tgz cacheable request tgz x normalize url tgz vulnerable library found in base branch master vulnerability details the normalize url package before x before and x before for node js has a redos regular expression denial of service issue because it has exponential performance for data urls publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution normalize url direct dependency fix resolution gatsby plugin sharp step up your open source security game with whitesource
0
122,856
12,167,373,648
IssuesEvent
2020-04-27 10:48:16
kubeflow/tf-operator
https://api.github.com/repos/kubeflow/tf-operator
closed
Update instructions and code to work with Kubernetes 1.8
kind/documentation lifecycle/stale
With Kubernetes 1.8 * GPU type has changed and is now ``` nvidia.com/gpu ``` * Volume mounting NVIDIA drivers from the host should no longer be necessary if the cluster use [device plugins](https://kubernetes.io/docs/concepts/cluster-administration/device-plugins/) @sozercan @wbuchwalter Is Azure using device plugins for GPUs or do you still rely on volume mounting GPU drivers?
1.0
Update instructions and code to work with Kubernetes 1.8 - With Kubernetes 1.8 * GPU type has changed and is now ``` nvidia.com/gpu ``` * Volume mounting NVIDIA drivers from the host should no longer be necessary if the cluster use [device plugins](https://kubernetes.io/docs/concepts/cluster-administration/device-plugins/) @sozercan @wbuchwalter Is Azure using device plugins for GPUs or do you still rely on volume mounting GPU drivers?
non_process
update instructions and code to work with kubernetes with kubernetes gpu type has changed and is now nvidia com gpu volume mounting nvidia drivers from the host should no longer be necessary if the cluster use sozercan wbuchwalter is azure using device plugins for gpus or do you still rely on volume mounting gpu drivers
0
7,311
10,449,982,029
IssuesEvent
2019-09-19 09:33:50
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Add chat system for traders communcation
in:gui in:trade-process
Once the trade has started the peers should have the ability to communicate with each other via a Bisq internal chat system (same as used in arbitration). It requires clear rules and the ability to ignore communication if user prefers. Any violation of the rules (e.g. trying social engineering scam) will lead to a dispute where the violating peer risks to lose his security deposit. With the messaging system in place it would be handy if the buyer could release the btc to the seller also, in case the buyer is unable to pay for some reason.
1.0
Add chat system for traders communcation - Once the trade has started the peers should have the ability to communicate with each other via a Bisq internal chat system (same as used in arbitration). It requires clear rules and the ability to ignore communication if user prefers. Any violation of the rules (e.g. trying social engineering scam) will lead to a dispute where the violating peer risks to lose his security deposit. With the messaging system in place it would be handy if the buyer could release the btc to the seller also, in case the buyer is unable to pay for some reason.
process
add chat system for traders communcation once the trade has started the peers should have the ability to communicate with each other via a bisq internal chat system same as used in arbitration it requires clear rules and the ability to ignore communication if user prefers any violation of the rules e g trying social engineering scam will lead to a dispute where the violating peer risks to lose his security deposit with the messaging system in place it would be handy if the buyer could release the btc to the seller also in case the buyer is unable to pay for some reason
1
9,451
12,429,236,838
IssuesEvent
2020-05-25 08:04:21
varys-main/ps-tools
https://api.github.com/repos/varys-main/ps-tools
opened
Docker - Anpassungen
processing
# User Story - Die Rechtschreibung soll angepasst werden und die Antworten auf Abfragen (J/N) sollen standardisiert werden. Zudem soll die Standard-Option besser ersichtlich sein. - Das Laden des Moduls NAV-ContainerHelper soll besser erklärt werden und/oder in eine eigene Funktion ausgelagert werden. Das Modul soll aktualisiert werden können (keine Installation mehrerer Versionen) - Die gu-Module sollen bereinigt werden. # Tasks - [ ] Rechtschreibung/Erklärung - [ ] NAV-ContainerHelper - [ ] gu-Module # Implementation # Known Problems
1.0
Docker - Anpassungen - # User Story - Die Rechtschreibung soll angepasst werden und die Antworten auf Abfragen (J/N) sollen standardisiert werden. Zudem soll die Standard-Option besser ersichtlich sein. - Das Laden des Moduls NAV-ContainerHelper soll besser erklärt werden und/oder in eine eigene Funktion ausgelagert werden. Das Modul soll aktualisiert werden können (keine Installation mehrerer Versionen) - Die gu-Module sollen bereinigt werden. # Tasks - [ ] Rechtschreibung/Erklärung - [ ] NAV-ContainerHelper - [ ] gu-Module # Implementation # Known Problems
process
docker anpassungen user story die rechtschreibung soll angepasst werden und die antworten auf abfragen j n sollen standardisiert werden zudem soll die standard option besser ersichtlich sein das laden des moduls nav containerhelper soll besser erklärt werden und oder in eine eigene funktion ausgelagert werden das modul soll aktualisiert werden können keine installation mehrerer versionen die gu module sollen bereinigt werden tasks rechtschreibung erklärung nav containerhelper gu module implementation known problems
1
722,124
24,850,850,600
IssuesEvent
2022-10-26 19:55:50
bounswe/bounswe2022group8
https://api.github.com/repos/bounswe/bounswe2022group8
opened
FE-4: Connection Between Backend and Frontend
Effort: High Priority: High Status: review needed coding frontend
### What's up? With the completion of backend part of login and signup, we have to connect backend to frontend by providing api calls on frontend and solving some authorization issues on frontend. To do it, first I have to configure backend files and database connection to backend on my local to test if connection is done successfully. ### To Do - [ ] Configure backend tools on local to test if api call works - [ ] Connect signup to backend by providing api calls on frontend - [ ] Connect login to backend by providing api calls on frontend ### Deadline 23.10.2022 @12.00 ### Additional Information _No response_ ### Reviewers @kostanya @mumcusena
1.0
FE-4: Connection Between Backend and Frontend - ### What's up? With the completion of backend part of login and signup, we have to connect backend to frontend by providing api calls on frontend and solving some authorization issues on frontend. To do it, first I have to configure backend files and database connection to backend on my local to test if connection is done successfully. ### To Do - [ ] Configure backend tools on local to test if api call works - [ ] Connect signup to backend by providing api calls on frontend - [ ] Connect login to backend by providing api calls on frontend ### Deadline 23.10.2022 @12.00 ### Additional Information _No response_ ### Reviewers @kostanya @mumcusena
non_process
fe connection between backend and frontend what s up with the completion of backend part of login and signup we have to connect backend to frontend by providing api calls on frontend and solving some authorization issues on frontend to do it first i have to configure backend files and database connection to backend on my local to test if connection is done successfully to do configure backend tools on local to test if api call works connect signup to backend by providing api calls on frontend connect login to backend by providing api calls on frontend deadline additional information no response reviewers kostanya mumcusena
0
15,781
27,884,245,655
IssuesEvent
2023-03-21 22:06:52
adrrf/acme-life-long-learning
https://api.github.com/repos/adrrf/acme-life-long-learning
opened
d3-s3-fr-16 - assistant operations on dashboards
develop functional_requirement
Operations by assistants on assistant dashboards: - [ ] Show their assistant dashboards.
1.0
d3-s3-fr-16 - assistant operations on dashboards - Operations by assistants on assistant dashboards: - [ ] Show their assistant dashboards.
non_process
fr assistant operations on dashboards operations by assistants on assistant dashboards show their assistant dashboards
0
648,354
21,183,748,402
IssuesEvent
2022-04-08 10:29:53
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
reopened
run.jobs.e2e_test: test_end_to_end failed
priority: p1 type: bug api: run samples flakybot: issue flakybot: flaky
Note: #7511 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: fe8dd4ce6953f40afa7e6b5b797b22e7f4daf546 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/fbd7f62e-c5f5-4259-b0df-d003ef8c6303), [Sponge](http://sponge2/fbd7f62e-c5f5-4259-b0df-d003ef8c6303) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/run/jobs/e2e_test.py", line 38, in setup_job subprocess.check_call( File "/usr/local/lib/python3.9/subprocess.py", line 373, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['gcloud', 'builds', 'submit', '--config', 'e2e_test_setup.yaml', '--project', 'python-docs-samples-tests', '--substitutions', '_SERVICE=job-quickstart-0cab165ab6,_VERSION=0cab165ab6']' returned non-zero exit status 1.</pre></details>
1.0
run.jobs.e2e_test: test_end_to_end failed - Note: #7511 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: fe8dd4ce6953f40afa7e6b5b797b22e7f4daf546 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/fbd7f62e-c5f5-4259-b0df-d003ef8c6303), [Sponge](http://sponge2/fbd7f62e-c5f5-4259-b0df-d003ef8c6303) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/run/jobs/e2e_test.py", line 38, in setup_job subprocess.check_call( File "/usr/local/lib/python3.9/subprocess.py", line 373, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['gcloud', 'builds', 'submit', '--config', 'e2e_test_setup.yaml', '--project', 'python-docs-samples-tests', '--substitutions', '_SERVICE=job-quickstart-0cab165ab6,_VERSION=0cab165ab6']' returned non-zero exit status 1.</pre></details>
non_process
run jobs test test end to end failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace run jobs test py line in setup job subprocess check call file usr local lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status
0
1,947
4,770,557,438
IssuesEvent
2016-10-26 15:33:56
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Rename url column to avoid conflicts in payload
API Explorer Processors
Currently `documents`, `files` and `sources` tables have a `url` column. This conflicts with our convention to send the OpenTrials API URL in the payload. Ex: ``` { total_count: 100, items: [ { id: 'document_id', url: 'http://api.opentrials.net/v1/documents/document_id', name: 'name', // ... } // ... ``` So we should rename the `url` column to `source_url` for `documents`, `files` and `sources`.
1.0
Rename url column to avoid conflicts in payload - Currently `documents`, `files` and `sources` tables have a `url` column. This conflicts with our convention to send the OpenTrials API URL in the payload. Ex: ``` { total_count: 100, items: [ { id: 'document_id', url: 'http://api.opentrials.net/v1/documents/document_id', name: 'name', // ... } // ... ``` So we should rename the `url` column to `source_url` for `documents`, `files` and `sources`.
process
rename url column to avoid conflicts in payload currently documents files and sources tables have a url column this conflicts with our convention to send the opentrials api url in the payload ex total count items id document id url name name so we should rename the url column to source url for documents files and sources
1
301,650
9,222,914,459
IssuesEvent
2019-03-12 01:01:27
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
opened
secure.meetup.com - design is broken
browser-firefox priority-important
<!-- @browser: Firefox 66.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:66.0) Gecko/20100101 Firefox/66.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://secure.meetup.com/login/ **Browser / Version**: Firefox 66.0 **Operating System**: Mac OS X 10.14 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Login with facebook won't work with firefox dev edition, works fine with google chrome **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190307095232</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-T2LNGD&l=gtmDataLayer was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://js-agent.newrelic.com/nr-spa-1118.min.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
secure.meetup.com - design is broken - <!-- @browser: Firefox 66.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:66.0) Gecko/20100101 Firefox/66.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://secure.meetup.com/login/ **Browser / Version**: Firefox 66.0 **Operating System**: Mac OS X 10.14 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Login with facebook won't work with firefox dev edition, works fine with google chrome **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190307095232</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: aurora</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-T2LNGD&l=gtmDataLayer was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignoring \'unsafe-inline\' within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring https: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "Content Security Policy: Ignoring http: within script-src: strict-dynamic specified"]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://js-agent.newrelic.com/nr-spa-1118.min.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.com/en_US/sdk.js was blocked because content blocking is enabled." {file: "https://secure.meetup.com/login/" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
secure meetup com design is broken url browser version firefox operating system mac os x tested another browser yes problem type design is broken description login with facebook won t work with firefox dev edition works fine with google chrome steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel aurora console messages u u u u u u u u u u u u u u u u u u from with ❤️
0
152,154
12,093,269,905
IssuesEvent
2020-04-19 18:57:00
gisellemartel/CONPASS
https://api.github.com/repos/gisellemartel/CONPASS
opened
AT-25 - (US5B - As a user, I want to get map directions to external points of interest.)
Acceptance Test (SPRINT 4)
1. Launch application 2. Select a building that supports indoor mode such as Hall and enter indoor mode by clicking on the building 3. Click on the direction button on the bottom right hand corner (blue button) 4. Enter a starting point (as you are inside) 5. Enter a destination outside like a local Tim Hortons
1.0
AT-25 - (US5B - As a user, I want to get map directions to external points of interest.) - 1. Launch application 2. Select a building that supports indoor mode such as Hall and enter indoor mode by clicking on the building 3. Click on the direction button on the bottom right hand corner (blue button) 4. Enter a starting point (as you are inside) 5. Enter a destination outside like a local Tim Hortons
non_process
at as a user i want to get map directions to external points of interest launch application select a building that supports indoor mode such as hall and enter indoor mode by clicking on the building click on the direction button on the bottom right hand corner blue button enter a starting point as you are inside enter a destination outside like a local tim hortons
0
16,693
21,792,169,331
IssuesEvent
2022-05-15 04:02:20
amor71/LiuAlgoTrader
https://api.github.com/repos/amor71/LiuAlgoTrader
closed
use contextvars
in-process no-issue-activity
Use python `contextvars` similar to how FastApi uses them to pass data to strategies and internally to simplify strategy development and code improvements. Need to POC to check the impact on performance
1.0
use contextvars - Use python `contextvars` similar to how FastApi uses them to pass data to strategies and internally to simplify strategy development and code improvements. Need to POC to check the impact on performance
process
use contextvars use python contextvars similar to how fastapi uses them to pass data to strategies and internally to simplify strategy development and code improvements need to poc to check the impact on performance
1
427,640
29,831,641,300
IssuesEvent
2023-06-18 10:52:09
io7m/medrina
https://api.github.com/repos/io7m/medrina
opened
Consider adjusting the syntax
documentation enhancement
Currently, rules look like this: ``` [allow-immediately [subject [with-any-roles inventory.admin]] [object true] [action true] ] ``` That's fine, but the problem it creates is that nothing actually says "this is a rule" explicitly. This means that it's hard to add new types of declarations later. This might be better: ``` [rule [allow-immediately [subject [with-any-roles inventory.admin]] [object true] [action true] ]] ``` This might also interact with #3 to yield something like: ``` [rule [description "Administrators are allowed to do anything."] [allow-immediately [subject [with-any-roles inventory.admin]] [object true] [action true] ]] ```
1.0
Consider adjusting the syntax - Currently, rules look like this: ``` [allow-immediately [subject [with-any-roles inventory.admin]] [object true] [action true] ] ``` That's fine, but the problem it creates is that nothing actually says "this is a rule" explicitly. This means that it's hard to add new types of declarations later. This might be better: ``` [rule [allow-immediately [subject [with-any-roles inventory.admin]] [object true] [action true] ]] ``` This might also interact with #3 to yield something like: ``` [rule [description "Administrators are allowed to do anything."] [allow-immediately [subject [with-any-roles inventory.admin]] [object true] [action true] ]] ```
non_process
consider adjusting the syntax currently rules look like this allow immediately that s fine but the problem it creates is that nothing actually says this is a rule explicitly this means that it s hard to add new types of declarations later this might be better rule allow immediately this might also interact with to yield something like rule allow immediately
0
114,228
9,693,445,938
IssuesEvent
2019-05-24 16:05:40
rubenverhoef/OAP-Config
https://api.github.com/repos/rubenverhoef/OAP-Config
opened
[HW] old screen cannot enable relay if it is on, only if it is off
Hardware HighPrio Needs Testing in Car bug
When the radio is on, it cannot enable the relay. I think the output which is used right now from the old screen is too weak. Related #45
1.0
[HW] old screen cannot enable relay if it is on, only if it is off - When the radio is on, it cannot enable the relay. I think the output which is used right now from the old screen is too weak. Related #45
non_process
old screen cannot enable relay if it is on only if it is off when the radio is on it cannot enable the relay i think the output which is used right now from the old screen is too weak related
0
139,210
31,280,660,641
IssuesEvent
2023-08-22 09:22:35
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Add an additional MatcherTransformType.CAMEL_KEEP_INTERNAL_CASING and PASCAL_KEEP_INTERNAL_CASING
T: Enhancement C: Code Generation P: Medium E: All Editions
The use case is: > For "passwordEncrypted", I want the generated getter to be "getPasswordEncrypted()". This is currently not possible. We would need a new `MatcherTransformType`, which transforms `a_bbbCd` into `aBbbCd` (`CAMEL_KEEP_INTERNAL_CASING`) or `ABbbCd` (`PASCAL_KEEP_INTERNAL_CASING`) --- See also this thread here: https://groups.google.com/d/msg/jooq-user/1iy0EdWe_T8/l7e3o_KTt-QJ
1.0
Add an additional MatcherTransformType.CAMEL_KEEP_INTERNAL_CASING and PASCAL_KEEP_INTERNAL_CASING - The use case is: > For "passwordEncrypted", I want the generated getter to be "getPasswordEncrypted()". This is currently not possible. We would need a new `MatcherTransformType`, which transforms `a_bbbCd` into `aBbbCd` (`CAMEL_KEEP_INTERNAL_CASING`) or `ABbbCd` (`PASCAL_KEEP_INTERNAL_CASING`) --- See also this thread here: https://groups.google.com/d/msg/jooq-user/1iy0EdWe_T8/l7e3o_KTt-QJ
non_process
add an additional matchertransformtype camel keep internal casing and pascal keep internal casing the use case is for passwordencrypted i want the generated getter to be getpasswordencrypted this is currently not possible we would need a new matchertransformtype which transforms a bbbcd into abbbcd camel keep internal casing or abbbcd pascal keep internal casing see also this thread here
0
17,709
23,602,549,975
IssuesEvent
2022-08-24 04:29:28
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Override general `IMessageRouter` methods to include Azure Service Bus request tracking
integration:service-bus area:message-processing telemetry
**Is your feature request related to a problem? Please describe.** Currently, we don't override the general message routing methods of the `IMessageRouter` interface (delegated via the general `MessageRouter`), whicih means that call via the general message routing methods won't result in Azure Service Bus telemetry requests. **Describe the solution you'd like** Override the general `RouteMessageAsync` method in the `AzureServiceBusMessageRouter` to include Azure Service Bus telemetry request tracking, like the other more Azure Service Bus-specific routing methods have.
1.0
Override general `IMessageRouter` methods to include Azure Service Bus request tracking - **Is your feature request related to a problem? Please describe.** Currently, we don't override the general message routing methods of the `IMessageRouter` interface (delegated via the general `MessageRouter`), whicih means that call via the general message routing methods won't result in Azure Service Bus telemetry requests. **Describe the solution you'd like** Override the general `RouteMessageAsync` method in the `AzureServiceBusMessageRouter` to include Azure Service Bus telemetry request tracking, like the other more Azure Service Bus-specific routing methods have.
process
override general imessagerouter methods to include azure service bus request tracking is your feature request related to a problem please describe currently we don t override the general message routing methods of the imessagerouter interface delegated via the general messagerouter whicih means that call via the general message routing methods won t result in azure service bus telemetry requests describe the solution you d like override the general routemessageasync method in the azureservicebusmessagerouter to include azure service bus telemetry request tracking like the other more azure service bus specific routing methods have
1
9,537
2,615,155,841
IssuesEvent
2015-03-01 06:34:00
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
Link at "overview on offline apps" in Client-Side Storage is broken
auto-migrated Priority-Medium Type-Defect
``` Link at "overview on offline apps" in "Summary" section of http://www.html5rocks.com/tutorials/offline/storage/ (Client-Side Storage) lead to 404 page I think it may link to http://www.html5rocks.com/tutorials/offline/whats-offline/ (missing / in target URL) ``` Original issue reported on code.google.com by `mementot...@gmail.com` on 19 Nov 2010 at 9:50
1.0
Link at "overview on offline apps" in Client-Side Storage is broken - ``` Link at "overview on offline apps" in "Summary" section of http://www.html5rocks.com/tutorials/offline/storage/ (Client-Side Storage) lead to 404 page I think it may link to http://www.html5rocks.com/tutorials/offline/whats-offline/ (missing / in target URL) ``` Original issue reported on code.google.com by `mementot...@gmail.com` on 19 Nov 2010 at 9:50
non_process
link at overview on offline apps in client side storage is broken link at overview on offline apps in summary section of client side storage lead to page i think it may link to missing in target url original issue reported on code google com by mementot gmail com on nov at
0
59,117
14,529,024,068
IssuesEvent
2020-12-14 17:16:14
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
iam.api-client.access_test: test_set_policy failed
api: iam buildcop: issue priority: p1 samples type: bug
Note: #3615 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 02abb311f20f81bf7b68b4064b7ceefe0f7018ed buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cb4a3071-d944-44cb-a238-f6a2298a23b3), [Sponge](http://sponge2/cb4a3071-d944-44cb-a238-f6a2298a23b3) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/iam/api-client/access_test.py", line 91, in test_set_policy test_call() File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 49, in wrapped_f return Retrying(*dargs, **dkw).call(f, *args, **kw) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 206, in call return attempt.get(self._wrap_exception) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 247, in get six.reraise(self.value[0], self.value[1], self.value[2]) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/six.py", line 703, in reraise raise value File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 200, in call attempt = Attempt(fn(*args, **kwargs), attempt_number, False) File "/workspace/iam/api-client/access_test.py", line 87, in test_call policy = access.get_policy(GCLOUD_PROJECT, version=3) File "/workspace/iam/api-client/access.py", line 40, in get_policy service.projects() File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/_helpers.py", line 134, in positional_wrapper return wrapped(*args, **kwargs) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 900, in execute resp, content = _retry_request( File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 177, in _retry_request resp, content = http.request(uri, method, *args, **kwargs) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google_auth_httplib2.py", line 189, in request self.credentials.before_request( File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/auth/credentials.py", line 133, in before_request self.refresh(request) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/service_account.py", line 361, in refresh access_token, expiry, _ = _client.jwt_grant(request, self._token_uri, assertion) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 153, in jwt_grant response_data = _token_endpoint_request(request, token_uri, body) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 111, in _token_endpoint_request response_data = json.loads(response_body) File "/usr/local/lib/python3.8/json/__init__.py", line 357, in loads return _default_decoder.decode(s) File "/usr/local/lib/python3.8/json/decoder.py", line 337, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/local/lib/python3.8/json/decoder.py", line 355, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)</pre></details>
1.0
iam.api-client.access_test: test_set_policy failed - Note: #3615 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 02abb311f20f81bf7b68b4064b7ceefe0f7018ed buildURL: [Build Status](https://source.cloud.google.com/results/invocations/cb4a3071-d944-44cb-a238-f6a2298a23b3), [Sponge](http://sponge2/cb4a3071-d944-44cb-a238-f6a2298a23b3) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/iam/api-client/access_test.py", line 91, in test_set_policy test_call() File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 49, in wrapped_f return Retrying(*dargs, **dkw).call(f, *args, **kw) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 206, in call return attempt.get(self._wrap_exception) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 247, in get six.reraise(self.value[0], self.value[1], self.value[2]) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/six.py", line 703, in reraise raise value File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/retrying.py", line 200, in call attempt = Attempt(fn(*args, **kwargs), attempt_number, False) File "/workspace/iam/api-client/access_test.py", line 87, in test_call policy = access.get_policy(GCLOUD_PROJECT, version=3) File "/workspace/iam/api-client/access.py", line 40, in get_policy service.projects() File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/_helpers.py", line 134, in positional_wrapper return wrapped(*args, **kwargs) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 900, in execute resp, content = _retry_request( File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/googleapiclient/http.py", line 177, in _retry_request resp, content = http.request(uri, method, *args, **kwargs) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google_auth_httplib2.py", line 189, in request self.credentials.before_request( File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/auth/credentials.py", line 133, in before_request self.refresh(request) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/service_account.py", line 361, in refresh access_token, expiry, _ = _client.jwt_grant(request, self._token_uri, assertion) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 153, in jwt_grant response_data = _token_endpoint_request(request, token_uri, body) File "/workspace/iam/api-client/.nox/py-3-8/lib/python3.8/site-packages/google/oauth2/_client.py", line 111, in _token_endpoint_request response_data = json.loads(response_body) File "/usr/local/lib/python3.8/json/__init__.py", line 357, in loads return _default_decoder.decode(s) File "/usr/local/lib/python3.8/json/decoder.py", line 337, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/local/lib/python3.8/json/decoder.py", line 355, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)</pre></details>
non_process
iam api client access test test set policy failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace iam api client access test py line in test set policy test call file workspace iam api client nox py lib site packages retrying py line in wrapped f return retrying dargs dkw call f args kw file workspace iam api client nox py lib site packages retrying py line in call return attempt get self wrap exception file workspace iam api client nox py lib site packages retrying py line in get six reraise self value self value self value file workspace iam api client nox py lib site packages six py line in reraise raise value file workspace iam api client nox py lib site packages retrying py line in call attempt attempt fn args kwargs attempt number false file workspace iam api client access test py line in test call policy access get policy gcloud project version file workspace iam api client access py line in get policy service projects file workspace iam api client nox py lib site packages googleapiclient helpers py line in positional wrapper return wrapped args kwargs file workspace iam api client nox py lib site packages googleapiclient http py line in execute resp content retry request file workspace iam api client nox py lib site packages googleapiclient http py line in retry request resp content http request uri method args kwargs file workspace iam api client nox py lib site packages google auth py line in request self credentials before request file workspace iam api client nox py lib site packages google auth credentials py line in before request self refresh request file workspace iam api client nox py lib site packages google service account py line in refresh access token expiry client jwt grant request self token uri assertion file workspace iam api client nox py lib site packages google client py line in jwt grant response data token endpoint request request token uri body file workspace iam api client nox py lib site packages google client py line in token endpoint request response data json loads response body file usr local lib json init py line in loads return default decoder decode s file usr local lib json decoder py line in decode obj end self raw decode s idx w s end file usr local lib json decoder py line in raw decode raise jsondecodeerror expecting value s err value from none json decoder jsondecodeerror expecting value line column char
0
66,985
14,814,070,712
IssuesEvent
2021-01-14 03:46:08
JoePep09/WebGoat
https://api.github.com/repos/JoePep09/WebGoat
opened
CVE-2018-20676 (Medium) detected in bootstrap-3.1.1.min.js
security vulnerability
## CVE-2018-20676 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.1.1.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js</a></p> <p>Path to vulnerable library: WebGoat/webgoat-lessons/challenge/src/main/resources/js/bootstrap.min.js,WebGoat/webgoat-container/target/classes/static/plugins/bootstrap/js/bootstrap.min.js,WebGoat/webgoat-container/src/main/resources/static/plugins/bootstrap/js/bootstrap.min.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/JoePep09/WebGoat/commit/cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2">cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20676 (Medium) detected in bootstrap-3.1.1.min.js - ## CVE-2018-20676 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.1.1.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.1.1/js/bootstrap.min.js</a></p> <p>Path to vulnerable library: WebGoat/webgoat-lessons/challenge/src/main/resources/js/bootstrap.min.js,WebGoat/webgoat-container/target/classes/static/plugins/bootstrap/js/bootstrap.min.js,WebGoat/webgoat-container/src/main/resources/static/plugins/bootstrap/js/bootstrap.min.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.1.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/JoePep09/WebGoat/commit/cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2">cd1b2f013f5ca37365f7a2990dd6a3bc2d7919d2</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library webgoat webgoat lessons challenge src main resources js bootstrap min js webgoat webgoat container target classes static plugins bootstrap js bootstrap min js webgoat webgoat container src main resources static plugins bootstrap js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch develop vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap step up your open source security game with whitesource
0
8,306
11,463,675,091
IssuesEvent
2020-02-07 16:27:12
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Solid Fuel display/database
Process Heating
![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/1d3e1c0a-8504-4550-bcc4-c0a9b562ef06) These should be showing up as 51.7%, 5.4%, 3.1%, 0%, 0.2%, 39.5%, and 0.1%. I think the tool is already accounting for them being in % format instead of fraction (hence "Difference" = 99%).
1.0
Solid Fuel display/database - ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/1d3e1c0a-8504-4550-bcc4-c0a9b562ef06) These should be showing up as 51.7%, 5.4%, 3.1%, 0%, 0.2%, 39.5%, and 0.1%. I think the tool is already accounting for them being in % format instead of fraction (hence "Difference" = 99%).
process
solid fuel display database these should be showing up as and i think the tool is already accounting for them being in format instead of fraction hence difference
1
4,277
7,190,483,476
IssuesEvent
2018-02-02 17:24:13
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
closed
ImageMagick 7 incompatibilities (was Rotation of an included graphics leads to error in latexmlpost
bug postprocessing
A simple example document with an included graphic like ``` \documentclass[12pt,a4paper,twoside]{book} \usepackage{graphicx} \begin{document} \begin{figure}[t] \includegraphics[angle=90,width=0.7\linewidth]{some.eps} \end{figure} \end{document} ``` leads to an error in the latexml post processing ``` $ latexml testgraphics.tex --dest=testgraphics.xml ... Conversion complete: No obvious problems. $ latexmlpost --format=html5 --dest=testgraphics.html testgraphics.xml ... (Graphics testgraphics.html processing... Error:imageprocessing:Rotate Image processing operation Rotate (degrees, -90, color, #FFFFFF) returned Exception 410: unrecognized option `color' @ error/Q16HDRI.xs/XS_Image__Magick__Q16HDRI_Mogrify/7807 In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic Warning:expected:image Couldn't get usable image for some.eps In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic 0.75 sec) (XSLT[using LaTeXML-html5.xsl] testgraphics.html processing... 0.02 sec) ``` The message ``Exception 410: unrecognized option `color'`` seems harmless to me. Maybe the error handling is a bit too strict. I think the processing of the image should not be aborted, because of this issue.
1.0
ImageMagick 7 incompatibilities (was Rotation of an included graphics leads to error in latexmlpost - A simple example document with an included graphic like ``` \documentclass[12pt,a4paper,twoside]{book} \usepackage{graphicx} \begin{document} \begin{figure}[t] \includegraphics[angle=90,width=0.7\linewidth]{some.eps} \end{figure} \end{document} ``` leads to an error in the latexml post processing ``` $ latexml testgraphics.tex --dest=testgraphics.xml ... Conversion complete: No obvious problems. $ latexmlpost --format=html5 --dest=testgraphics.html testgraphics.xml ... (Graphics testgraphics.html processing... Error:imageprocessing:Rotate Image processing operation Rotate (degrees, -90, color, #FFFFFF) returned Exception 410: unrecognized option `color' @ error/Q16HDRI.xs/XS_Image__Magick__Q16HDRI_Mogrify/7807 In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic Warning:expected:image Couldn't get usable image for some.eps In Post::Graphics[@0x7fb0e3bca0f8] ->transformGraphic 0.75 sec) (XSLT[using LaTeXML-html5.xsl] testgraphics.html processing... 0.02 sec) ``` The message ``Exception 410: unrecognized option `color'`` seems harmless to me. Maybe the error handling is a bit too strict. I think the processing of the image should not be aborted, because of this issue.
process
imagemagick incompatibilities was rotation of an included graphics leads to error in latexmlpost a simple example document with an included graphic like documentclass book usepackage graphicx begin document begin figure includegraphics some eps end figure end document leads to an error in the latexml post processing latexml testgraphics tex dest testgraphics xml conversion complete no obvious problems latexmlpost format dest testgraphics html testgraphics xml graphics testgraphics html processing error imageprocessing rotate image processing operation rotate degrees color ffffff returned exception unrecognized option color error xs xs image magick mogrify in post graphics transformgraphic warning expected image couldn t get usable image for some eps in post graphics transformgraphic sec xslt testgraphics html processing sec the message exception unrecognized option color seems harmless to me maybe the error handling is a bit too strict i think the processing of the image should not be aborted because of this issue
1
63,275
26,329,935,191
IssuesEvent
2023-01-10 10:01:49
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
CognitiveServices Accounts are not able to use the list-skus feature in the az cli
question Cognitive Services customer-reported CXP Attention Auto-Assign issue-addressed
<!--- 🛑 Please check existing issues first before continuing: https://github.com/Azure/azure-cli/issues ---> ### **This is autogenerated. Please review and update as needed.** ## Describe the bug **Command Name** `az cognitiveservices account list-skus` **Errors:** ``` The command failed with an unexpected error. Here is the traceback: No value for given attribute Traceback (most recent call last): File "/usr/lib64/az/lib/python3.9/site-packages/knack/cli.py", line 233, in invoke cmd_result = self.invocation.execute(args) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 663, in execute raise ex File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 726, in _run_jobs_serially results.append(self._run_job(expanded_arg, cmd_copy)) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 697, in _run_job result = cmd_copy(params) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 333, in __call__ return self.handler(*args, **kwargs) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/command_operation.py", line 121, in handler return op(**command_args) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/command_modules/cognitiveservices/custom.py", line 75, in list_skus return cf_accounts(cmd.cli_ctx).list_skus(resource_group_name, account_name) File "/usr/lib64/az/lib/python3.9/site-packages/azure/core/tracing/decorator.py", line 73, in wrapper_use_tracer return func(*args, **kwargs) File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 1354, in list_skus request = build_list_skus_request( File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 330, in build_list_skus_request "accountName": _SERIALIZER.url( File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 703, in url output = self.serialize_data(data, data_type, **kwargs) File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 780, in serialize_data raise ValueError("No value for given attribute") ValueError: No value for given attribute ``` ## To Reproduce: Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information. - _Put any pre-requisite steps here..._ - `az cognitiveservices account list-skus --kind {}` ## Expected Behavior ## Environment Summary ``` Linux-5.4.0-1098-azure-x86_64-with-glibc2.35 (Cloud Shell), CBL-Mariner/Linux Python 3.9.14 Installer: RPM azure-cli 2.43.0 Extensions: quota 0.1.0 connectedk8s 1.3.3 ai-examples 0.2.5 ml 2.11.0 ssh 1.1.3 Dependencies: msal 1.20.0 azure-mgmt-resource 21.1.0b1 ``` ## Additional Context <!--Please don't remove this:--> <!--auto-generated-->
1.0
CognitiveServices Accounts are not able to use the list-skus feature in the az cli - <!--- 🛑 Please check existing issues first before continuing: https://github.com/Azure/azure-cli/issues ---> ### **This is autogenerated. Please review and update as needed.** ## Describe the bug **Command Name** `az cognitiveservices account list-skus` **Errors:** ``` The command failed with an unexpected error. Here is the traceback: No value for given attribute Traceback (most recent call last): File "/usr/lib64/az/lib/python3.9/site-packages/knack/cli.py", line 233, in invoke cmd_result = self.invocation.execute(args) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 663, in execute raise ex File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 726, in _run_jobs_serially results.append(self._run_job(expanded_arg, cmd_copy)) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 697, in _run_job result = cmd_copy(params) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/__init__.py", line 333, in __call__ return self.handler(*args, **kwargs) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/core/commands/command_operation.py", line 121, in handler return op(**command_args) File "/usr/lib64/az/lib/python3.9/site-packages/azure/cli/command_modules/cognitiveservices/custom.py", line 75, in list_skus return cf_accounts(cmd.cli_ctx).list_skus(resource_group_name, account_name) File "/usr/lib64/az/lib/python3.9/site-packages/azure/core/tracing/decorator.py", line 73, in wrapper_use_tracer return func(*args, **kwargs) File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 1354, in list_skus request = build_list_skus_request( File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/operations/_accounts_operations.py", line 330, in build_list_skus_request "accountName": _SERIALIZER.url( File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 703, in url output = self.serialize_data(data, data_type, **kwargs) File "/usr/lib64/az/lib/python3.9/site-packages/azure/mgmt/cognitiveservices/_serialization.py", line 780, in serialize_data raise ValueError("No value for given attribute") ValueError: No value for given attribute ``` ## To Reproduce: Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information. - _Put any pre-requisite steps here..._ - `az cognitiveservices account list-skus --kind {}` ## Expected Behavior ## Environment Summary ``` Linux-5.4.0-1098-azure-x86_64-with-glibc2.35 (Cloud Shell), CBL-Mariner/Linux Python 3.9.14 Installer: RPM azure-cli 2.43.0 Extensions: quota 0.1.0 connectedk8s 1.3.3 ai-examples 0.2.5 ml 2.11.0 ssh 1.1.3 Dependencies: msal 1.20.0 azure-mgmt-resource 21.1.0b1 ``` ## Additional Context <!--Please don't remove this:--> <!--auto-generated-->
non_process
cognitiveservices accounts are not able to use the list skus feature in the az cli this is autogenerated please review and update as needed describe the bug command name az cognitiveservices account list skus errors the command failed with an unexpected error here is the traceback no value for given attribute traceback most recent call last file usr az lib site packages knack cli py line in invoke cmd result self invocation execute args file usr az lib site packages azure cli core commands init py line in execute raise ex file usr az lib site packages azure cli core commands init py line in run jobs serially results append self run job expanded arg cmd copy file usr az lib site packages azure cli core commands init py line in run job result cmd copy params file usr az lib site packages azure cli core commands init py line in call return self handler args kwargs file usr az lib site packages azure cli core commands command operation py line in handler return op command args file usr az lib site packages azure cli command modules cognitiveservices custom py line in list skus return cf accounts cmd cli ctx list skus resource group name account name file usr az lib site packages azure core tracing decorator py line in wrapper use tracer return func args kwargs file usr az lib site packages azure mgmt cognitiveservices operations accounts operations py line in list skus request build list skus request file usr az lib site packages azure mgmt cognitiveservices operations accounts operations py line in build list skus request accountname serializer url file usr az lib site packages azure mgmt cognitiveservices serialization py line in url output self serialize data data data type kwargs file usr az lib site packages azure mgmt cognitiveservices serialization py line in serialize data raise valueerror no value for given attribute valueerror no value for given attribute to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az cognitiveservices account list skus kind expected behavior environment summary linux azure with cloud shell cbl mariner linux python installer rpm azure cli extensions quota ai examples ml ssh dependencies msal azure mgmt resource additional context
0
248,899
7,940,735,128
IssuesEvent
2018-07-10 00:06:57
DLR-SC/tigl
https://api.github.com/repos/DLR-SC/tigl
closed
Inconsistent API CCPACSFuselage::GetLoft
Priority-High
The method CCPACSFuselage::GetLoft is inconsistent in two ways since commit fb87cd650bc39: 1. It returns a TopoDS_Shape instead of a PNamedShape 2. It has a transformation argument. The default API doen not have this argument. This is problematic for e.g. the Python API since it overwrites the old behavior and the python examples don't work.
1.0
Inconsistent API CCPACSFuselage::GetLoft - The method CCPACSFuselage::GetLoft is inconsistent in two ways since commit fb87cd650bc39: 1. It returns a TopoDS_Shape instead of a PNamedShape 2. It has a transformation argument. The default API doen not have this argument. This is problematic for e.g. the Python API since it overwrites the old behavior and the python examples don't work.
non_process
inconsistent api ccpacsfuselage getloft the method ccpacsfuselage getloft is inconsistent in two ways since commit it returns a topods shape instead of a pnamedshape it has a transformation argument the default api doen not have this argument this is problematic for e g the python api since it overwrites the old behavior and the python examples don t work
0
15,997
2,870,250,336
IssuesEvent
2015-06-07 00:34:14
pdelia/away3d
https://api.github.com/repos/pdelia/away3d
closed
Rendering problem with TextField3D and Trident
auto-migrated Priority-Medium Type-Defect
#49 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:23Z ``` What steps will reproduce the problem? 1. Add TextField3D and Trident to same Away3D scene 2. Render. What is the expected output? What do you see instead? In addition to the 3D text and the trident, I see additional spurious lines and shapes. What version of the product are you using? On what operating system? Away3D trunk rev 1404 (17 June 2009). Flash CS3. Windows XP. Please provide any additional information below. FLA demonstrating the problem is attached. Same problem seems to occur if I mix a Sprite2D with a Trident or any object using LineSegment. ``` Original issue reported on code.google.com by `ralph%sa...@gtempaccount.com` on 17 Jun 2009 at 11:20 Attachments: * [renderBug.fla](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.fla) * [renderBug.swf](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.swf)
1.0
Rendering problem with TextField3D and Trident - #49 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:23Z ``` What steps will reproduce the problem? 1. Add TextField3D and Trident to same Away3D scene 2. Render. What is the expected output? What do you see instead? In addition to the 3D text and the trident, I see additional spurious lines and shapes. What version of the product are you using? On what operating system? Away3D trunk rev 1404 (17 June 2009). Flash CS3. Windows XP. Please provide any additional information below. FLA demonstrating the problem is attached. Same problem seems to occur if I mix a Sprite2D with a Trident or any object using LineSegment. ``` Original issue reported on code.google.com by `ralph%sa...@gtempaccount.com` on 17 Jun 2009 at 11:20 Attachments: * [renderBug.fla](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.fla) * [renderBug.swf](https://storage.googleapis.com/google-code-attachments/away3d/issue-49/comment-0/renderBug.swf)
non_process
rendering problem with and trident issue by googlecodeexporter created on what steps will reproduce the problem add and trident to same scene render what is the expected output what do you see instead in addition to the text and the trident i see additional spurious lines and shapes what version of the product are you using on what operating system trunk rev june flash windows xp please provide any additional information below fla demonstrating the problem is attached same problem seems to occur if i mix a with a trident or any object using linesegment original issue reported on code google com by ralph sa gtempaccount com on jun at attachments
0
89,224
3,790,766,636
IssuesEvent
2016-03-21 22:48:12
projectcalico/calico-containers
https://api.github.com/repos/projectcalico/calico-containers
closed
Manual setup fails to connect to bird
Priority: Medium
Running the manual, two VM setup for a trial on Centos. Docker container up, calico containers fire up and I can't seem to connect to bird socket. $ sudo /home/cloud-user/calicoctl node --libnetwork No IP provided. Using detected IP: 192.168.98.6 Calico node is running with id: 5277d2ec28e06e7a6bf336b9ce318a4355afd847bd648473c3f0b228cd89c169 Calico libnetwork driver is running with id: 9bc2c26970c1e91523a0ec394046d209a48d4bfe2d9bbf1ef96bbbd2fbec1427 [cloud-user@ken1 ~]$ calicoctl status calico-node container is running. Status: Up 12 seconds Running felix version 1.3.0rc6 IPv4 BGP status IP: 192.168.98.6 AS Number: 64511 (inherited) Unable to connect to server control socket (/var/run/calico/bird.ctl): Permission denied Couldn't connect to bird. Try running as root. IPv6 BGP status No IPv6 address configured.
1.0
Manual setup fails to connect to bird - Running the manual, two VM setup for a trial on Centos. Docker container up, calico containers fire up and I can't seem to connect to bird socket. $ sudo /home/cloud-user/calicoctl node --libnetwork No IP provided. Using detected IP: 192.168.98.6 Calico node is running with id: 5277d2ec28e06e7a6bf336b9ce318a4355afd847bd648473c3f0b228cd89c169 Calico libnetwork driver is running with id: 9bc2c26970c1e91523a0ec394046d209a48d4bfe2d9bbf1ef96bbbd2fbec1427 [cloud-user@ken1 ~]$ calicoctl status calico-node container is running. Status: Up 12 seconds Running felix version 1.3.0rc6 IPv4 BGP status IP: 192.168.98.6 AS Number: 64511 (inherited) Unable to connect to server control socket (/var/run/calico/bird.ctl): Permission denied Couldn't connect to bird. Try running as root. IPv6 BGP status No IPv6 address configured.
non_process
manual setup fails to connect to bird running the manual two vm setup for a trial on centos docker container up calico containers fire up and i can t seem to connect to bird socket sudo home cloud user calicoctl node libnetwork no ip provided using detected ip calico node is running with id calico libnetwork driver is running with id calicoctl status calico node container is running status up seconds running felix version bgp status ip as number inherited unable to connect to server control socket var run calico bird ctl permission denied couldn t connect to bird try running as root bgp status no address configured
0
301,648
26,082,932,690
IssuesEvent
2022-12-25 17:06:03
Sun-Mountain/lettuceMeetApp
https://api.github.com/repos/Sun-Mountain/lettuceMeetApp
closed
Test - Finish tests for registration and login
tests
Describe the task here and link or add dependencies to any applicable issues... ### This task is done when... - [ ] criteria 1 - [ ] criteria 2 --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
1.0
Test - Finish tests for registration and login - Describe the task here and link or add dependencies to any applicable issues... ### This task is done when... - [ ] criteria 1 - [ ] criteria 2 --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
non_process
test finish tests for registration and login describe the task here and link or add dependencies to any applicable issues this task is done when criteria criteria add additional labels design dev compliance bug etc and size before submitting if the issue is needed to complete prioritized work for the current sprint add it to the this sprint pipeline otherwise all other issues will be automatically added to the unprioritized pipeline for prioritization in backlog refinement or sprint planning with product
0
17,269
23,051,022,635
IssuesEvent
2022-07-24 16:24:56
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Music Hell from “We are Little Zombies” (Screenshots and Title Card Added)
suggested title in process
Please add as much of the following info as you can: Title: Music Hell Type (film/tv show): TV show - entertainment commentary (music like MTV shows) Film or show in which it appears: We Are Little Zombies Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? probably about halfway Actual footage of the film/show can be seen (yes/no)? Yes - You can see a little bit of it in the trailer. It leads into the band, Little Zombies', music video. https://youtu.be/obstTJVi3cE?t=74
1.0
Add Music Hell from “We are Little Zombies” (Screenshots and Title Card Added) - Please add as much of the following info as you can: Title: Music Hell Type (film/tv show): TV show - entertainment commentary (music like MTV shows) Film or show in which it appears: We Are Little Zombies Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? probably about halfway Actual footage of the film/show can be seen (yes/no)? Yes - You can see a little bit of it in the trailer. It leads into the band, Little Zombies', music video. https://youtu.be/obstTJVi3cE?t=74
process
add music hell from “we are little zombies” screenshots and title card added please add as much of the following info as you can title music hell type film tv show tv show entertainment commentary music like mtv shows film or show in which it appears we are little zombies is the parent film show streaming anywhere yes amazon prime about when in the parent film show does it appear probably about halfway actual footage of the film show can be seen yes no yes you can see a little bit of it in the trailer it leads into the band little zombies music video
1
8,578
11,746,976,375
IssuesEvent
2020-03-12 12:49:37
pwittchen/InfiniteScroll
https://api.github.com/repos/pwittchen/InfiniteScroll
opened
release 0.0.3
release process
**Release notes**: - updated project dependencies - updated build configuration **Things to do**: - [ ] bump library version - [ ] upload Archives to Maven Central Repository - [ ] close and release artifact on Nexus - [ ] update gh-pages with JavaDoc - [ ] update `CHANGELOG.md` after Maven Sync - [ ] update Download section in `README.md` after Maven Sync - [ ] create GitHub release
1.0
release 0.0.3 - **Release notes**: - updated project dependencies - updated build configuration **Things to do**: - [ ] bump library version - [ ] upload Archives to Maven Central Repository - [ ] close and release artifact on Nexus - [ ] update gh-pages with JavaDoc - [ ] update `CHANGELOG.md` after Maven Sync - [ ] update Download section in `README.md` after Maven Sync - [ ] create GitHub release
process
release release notes updated project dependencies updated build configuration things to do bump library version upload archives to maven central repository close and release artifact on nexus update gh pages with javadoc update changelog md after maven sync update download section in readme md after maven sync create github release
1
210,073
7,182,970,005
IssuesEvent
2018-02-01 11:40:31
crowdAI/crowdai
https://api.github.com/repos/crowdAI/crowdai
opened
Gitlab cert
high priority
Your certificate (or certificates) for the names listed below will expire in 20 days (on 21 Feb 18 11:46 +0000). Please make sure to renew your certificate before then, or visitors to your website will encounter errors. gitlab.crowdai.org
1.0
Gitlab cert - Your certificate (or certificates) for the names listed below will expire in 20 days (on 21 Feb 18 11:46 +0000). Please make sure to renew your certificate before then, or visitors to your website will encounter errors. gitlab.crowdai.org
non_process
gitlab cert your certificate or certificates for the names listed below will expire in days on feb please make sure to renew your certificate before then or visitors to your website will encounter errors gitlab crowdai org
0
10,072
14,444,468,965
IssuesEvent
2020-12-07 21:17:49
spacetelescope/jwst
https://api.github.com/repos/spacetelescope/jwst
closed
exposures that get Level 2b processing
Requirement associations
Prompted by [this comment](https://github.com/STScI-JWST/jwst/issues/735#issuecomment-292226126), this issue documents exactly which exposures will go through any type of Level 2b processing, such as Spec2Pipeline or Image2Pipeline. Please review, update, and if updated, leave a comment that one has done so. Questions in following comments. EXP_TYPE | Association Exposure Type | Specials | Level2b? ---------- | --------------------------- | --------- | --------- NRC_IMAGE | science | n/a | YES NRC_GRISM/NRC_WFSS | science | n/a | YES NRC_TACQ | target_acquisition | n/a | YES NRC_TACONFIRM | target_acquisition | n/a | YES NRC_CORON | science | n/a | YES NRC_CORON | psf | PSF | YES NRC_TSIMAGE | science | n/a | YES NRC_TSGRISM | science | n/a | YES NRC_FOCUS | science | n/a/ | YES NRC_DARK | dark | n/a | NO NRC_FLAT | flat | n/a | NO NRC_LED | science | n/a | NO | | | MIR_IMAGE | science | n/a | YES MIR_TACQ | target_acquisition | n/a | YES MIR_LYOT | science | n/a | YES MIR_LYOT | psf | PSF | YES MIR_4QPM | science | n/a | YES MIR_4QPM | psf | PSF | YES MIR_LRS-FIXEDSLIT | science | n/a | YES MIR_LRS-FIXEDSLIT | background | BACKGROUND | YES MIR_LRS-SLITLESS | science | n/a | YES MIR_LRS-SLITLESS | background | BACKGROUND | YES MIR_MRS | science | n/a | YES MIR_MRS | background | BACKGROUND | YES MIR_DARKIMG | dark | n/a | NO MIR_DARKMRS | dark | n/a | NO MIR_FLATIMAGE | flat | n/a | NO MIR_FLATIMAGE-EXT | flat | n/a | NO MIR_FLATMRS | flat | n/a | NO MIR_FLATMRS-EXT | flat | n/a | NO MIR_CORONCAL | science | n/a | YES | | | NRS_WATA | target_acquisition | n/a | YES NRS_MSATA | target_acquisition | n/a | YES NRS_TACONFIRM | target_acquisition | n/a | YES NRS_CONFIRM | science | n/a | YES NRS_FIXEDSLIT | science | n/a | YES NRS_FIXEDSLIT | background | BACKGROUND | YES NRS_AUTOWAVE | nrs_autowave | n/a | YES NRS_IFU | science | n/a | YES NRS_IFU | imprint | IMPRINT | YES NRS_IFU | background | BACKGROUND | YES NRS_IMAGE | science | n/a | YES NRS_MSASPEC | science | n/a | YES NRS_MSASPEC | imprint | IMPRINT | YES NRS_AUTOFLAT | nrs_autoflat | n/a | YES NRS_FOCUS | science | n/a | YES NRS_DARK | dark | n/a | NO NRS_LAMP | science | n/a | YES<sup>1</sup> NRS_BRIGHTOBJ | science | n/a | YES NRS_MIMF | science | n/a | YES NRS_VERIFY | science | n/a | YES<sup>1</sup> | | | NIS_IMAGE | science | n/a | YES NIS_WFSS | science | n/a | YES NIS_TACQ | target_acquisition | n/a | YES NIS_TACONFIRM | target_acquisition | n/a | YES NIS_SOSS | science | n/a | YES NIS_AMI | science | n/a | YES NIS_AMI | psf | PSF | YES NIS_FOCUS | science | n/a | YES NIS_DARK | science | n/a | NO NIS_LAMP | science | n/a | NO NIS_EXTCAL | science | n/a | NO | | | FGS_IMAGE | science | n/a | YES FGS_FOCUS | science | n/a | YES FGS_SKYFLAT | flat | n/a | NO FGS_INTFLAT | flat | n/a | NO FGS_DARK | dark | n/a | NO FGS_ID-STACK | tracking | n/a | NO FGS_ID-IMAGE | tracking | n/a | NO FGS_ACQ1 | tracking | n/a | NO FGS_ACQ2 | tracking | n/a | NO FGS_TRACK | tracking | n/a | NO FGS_FINEGUIDE | tracking | n/a | NO ## Notes <sup>1</sup> Custom Level2 processing
1.0
exposures that get Level 2b processing - Prompted by [this comment](https://github.com/STScI-JWST/jwst/issues/735#issuecomment-292226126), this issue documents exactly which exposures will go through any type of Level 2b processing, such as Spec2Pipeline or Image2Pipeline. Please review, update, and if updated, leave a comment that one has done so. Questions in following comments. EXP_TYPE | Association Exposure Type | Specials | Level2b? ---------- | --------------------------- | --------- | --------- NRC_IMAGE | science | n/a | YES NRC_GRISM/NRC_WFSS | science | n/a | YES NRC_TACQ | target_acquisition | n/a | YES NRC_TACONFIRM | target_acquisition | n/a | YES NRC_CORON | science | n/a | YES NRC_CORON | psf | PSF | YES NRC_TSIMAGE | science | n/a | YES NRC_TSGRISM | science | n/a | YES NRC_FOCUS | science | n/a/ | YES NRC_DARK | dark | n/a | NO NRC_FLAT | flat | n/a | NO NRC_LED | science | n/a | NO | | | MIR_IMAGE | science | n/a | YES MIR_TACQ | target_acquisition | n/a | YES MIR_LYOT | science | n/a | YES MIR_LYOT | psf | PSF | YES MIR_4QPM | science | n/a | YES MIR_4QPM | psf | PSF | YES MIR_LRS-FIXEDSLIT | science | n/a | YES MIR_LRS-FIXEDSLIT | background | BACKGROUND | YES MIR_LRS-SLITLESS | science | n/a | YES MIR_LRS-SLITLESS | background | BACKGROUND | YES MIR_MRS | science | n/a | YES MIR_MRS | background | BACKGROUND | YES MIR_DARKIMG | dark | n/a | NO MIR_DARKMRS | dark | n/a | NO MIR_FLATIMAGE | flat | n/a | NO MIR_FLATIMAGE-EXT | flat | n/a | NO MIR_FLATMRS | flat | n/a | NO MIR_FLATMRS-EXT | flat | n/a | NO MIR_CORONCAL | science | n/a | YES | | | NRS_WATA | target_acquisition | n/a | YES NRS_MSATA | target_acquisition | n/a | YES NRS_TACONFIRM | target_acquisition | n/a | YES NRS_CONFIRM | science | n/a | YES NRS_FIXEDSLIT | science | n/a | YES NRS_FIXEDSLIT | background | BACKGROUND | YES NRS_AUTOWAVE | nrs_autowave | n/a | YES NRS_IFU | science | n/a | YES NRS_IFU | imprint | IMPRINT | YES NRS_IFU | background | BACKGROUND | YES NRS_IMAGE | science | n/a | YES NRS_MSASPEC | science | n/a | YES NRS_MSASPEC | imprint | IMPRINT | YES NRS_AUTOFLAT | nrs_autoflat | n/a | YES NRS_FOCUS | science | n/a | YES NRS_DARK | dark | n/a | NO NRS_LAMP | science | n/a | YES<sup>1</sup> NRS_BRIGHTOBJ | science | n/a | YES NRS_MIMF | science | n/a | YES NRS_VERIFY | science | n/a | YES<sup>1</sup> | | | NIS_IMAGE | science | n/a | YES NIS_WFSS | science | n/a | YES NIS_TACQ | target_acquisition | n/a | YES NIS_TACONFIRM | target_acquisition | n/a | YES NIS_SOSS | science | n/a | YES NIS_AMI | science | n/a | YES NIS_AMI | psf | PSF | YES NIS_FOCUS | science | n/a | YES NIS_DARK | science | n/a | NO NIS_LAMP | science | n/a | NO NIS_EXTCAL | science | n/a | NO | | | FGS_IMAGE | science | n/a | YES FGS_FOCUS | science | n/a | YES FGS_SKYFLAT | flat | n/a | NO FGS_INTFLAT | flat | n/a | NO FGS_DARK | dark | n/a | NO FGS_ID-STACK | tracking | n/a | NO FGS_ID-IMAGE | tracking | n/a | NO FGS_ACQ1 | tracking | n/a | NO FGS_ACQ2 | tracking | n/a | NO FGS_TRACK | tracking | n/a | NO FGS_FINEGUIDE | tracking | n/a | NO ## Notes <sup>1</sup> Custom Level2 processing
non_process
exposures that get level processing prompted by this issue documents exactly which exposures will go through any type of level processing such as or please review update and if updated leave a comment that one has done so questions in following comments exp type association exposure type specials nrc image science n a yes nrc grism nrc wfss science n a yes nrc tacq target acquisition n a yes nrc taconfirm target acquisition n a yes nrc coron science n a yes nrc coron psf psf yes nrc tsimage science n a yes nrc tsgrism science n a yes nrc focus science n a yes nrc dark dark n a no nrc flat flat n a no nrc led science n a no mir image science n a yes mir tacq target acquisition n a yes mir lyot science n a yes mir lyot psf psf yes mir science n a yes mir psf psf yes mir lrs fixedslit science n a yes mir lrs fixedslit background background yes mir lrs slitless science n a yes mir lrs slitless background background yes mir mrs science n a yes mir mrs background background yes mir darkimg dark n a no mir darkmrs dark n a no mir flatimage flat n a no mir flatimage ext flat n a no mir flatmrs flat n a no mir flatmrs ext flat n a no mir coroncal science n a yes nrs wata target acquisition n a yes nrs msata target acquisition n a yes nrs taconfirm target acquisition n a yes nrs confirm science n a yes nrs fixedslit science n a yes nrs fixedslit background background yes nrs autowave nrs autowave n a yes nrs ifu science n a yes nrs ifu imprint imprint yes nrs ifu background background yes nrs image science n a yes nrs msaspec science n a yes nrs msaspec imprint imprint yes nrs autoflat nrs autoflat n a yes nrs focus science n a yes nrs dark dark n a no nrs lamp science n a yes nrs brightobj science n a yes nrs mimf science n a yes nrs verify science n a yes nis image science n a yes nis wfss science n a yes nis tacq target acquisition n a yes nis taconfirm target acquisition n a yes nis soss science n a yes nis ami science n a yes nis ami psf psf yes nis focus science n a yes nis dark science n a no nis lamp science n a no nis extcal science n a no fgs image science n a yes fgs focus science n a yes fgs skyflat flat n a no fgs intflat flat n a no fgs dark dark n a no fgs id stack tracking n a no fgs id image tracking n a no fgs tracking n a no fgs tracking n a no fgs track tracking n a no fgs fineguide tracking n a no notes custom processing
0
8,745
11,872,370,203
IssuesEvent
2020-03-26 15:43:10
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Othermeta and source metadata should not cascade to child topicref
bug preprocess priority/high
Here is an example of the present behaviour: The map: ``` xml <bookmap> <booktitle> <mainbooktitle>Simple example</mainbooktitle> </booktitle> <bookmeta> <othermeta name="this_meta" content="should not be inherited"/> </bookmeta> <chapter href="child.dita"/> </bookmap> ``` The topic: ``` xml <topic id="topic-1"> <title>The child</title> <body> <p>Should not have any meta.</p> </body> </topic> ``` The result xhtml transformation of the topic: ``` xml <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> <html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en-us" lang="en-us"> <head><meta http-equiv="Content-Type" content="text/html; charset=UTF-8" /> <meta name="copyright" content="(C) Copyright 2015" /> <meta name="DC.rights.owner" content="(C) Copyright 2015" /> <meta name="DC.Type" content="topic" /> <meta name="DC.Title" content="The child" /> <meta name="this_meta" content="should not be inherited" /> <meta name="DC.Format" content="XHTML" /> <meta name="DC.Identifier" content="topic-1" /> <link rel="stylesheet" type="text/css" href="commonltr.css" /> <title>The child</title> </head> <body id="topic-1"> <h1 class="title topictitle1" id="ariaid-title1">The child</h1> <div class="body"> <p class="p">Should not have any meta.</p> </div> </body> </html> ``` As you can see, the othermeta has been inherited by the topic. But, as stated in the [dita 1.2 spec](https://dita.fluidtopics.com/book#!book;uri=urn%253Adita%253ADITA1%255C0dita-1.2-specification.ditamap%253Fv%255C2DITA1.2-spec-XHTML;breadcrumb=urn%253Adita%253ADITA1%255C0architectural_specification.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C3-base.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0ditamarkup.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0dita%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0reconciling%255C3-topic%255C3-and%255C3-map%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML), othermeta and source metadata (among others) should not cascade to children. This bug has been introduced has part of the fix for #1086. The fix may be to revert the incriminated commit (the one that I am unable to find).
1.0
Othermeta and source metadata should not cascade to child topicref - Here is an example of the present behaviour: The map: ``` xml <bookmap> <booktitle> <mainbooktitle>Simple example</mainbooktitle> </booktitle> <bookmeta> <othermeta name="this_meta" content="should not be inherited"/> </bookmeta> <chapter href="child.dita"/> </bookmap> ``` The topic: ``` xml <topic id="topic-1"> <title>The child</title> <body> <p>Should not have any meta.</p> </body> </topic> ``` The result xhtml transformation of the topic: ``` xml <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> <html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en-us" lang="en-us"> <head><meta http-equiv="Content-Type" content="text/html; charset=UTF-8" /> <meta name="copyright" content="(C) Copyright 2015" /> <meta name="DC.rights.owner" content="(C) Copyright 2015" /> <meta name="DC.Type" content="topic" /> <meta name="DC.Title" content="The child" /> <meta name="this_meta" content="should not be inherited" /> <meta name="DC.Format" content="XHTML" /> <meta name="DC.Identifier" content="topic-1" /> <link rel="stylesheet" type="text/css" href="commonltr.css" /> <title>The child</title> </head> <body id="topic-1"> <h1 class="title topictitle1" id="ariaid-title1">The child</h1> <div class="body"> <p class="p">Should not have any meta.</p> </div> </body> </html> ``` As you can see, the othermeta has been inherited by the topic. But, as stated in the [dita 1.2 spec](https://dita.fluidtopics.com/book#!book;uri=urn%253Adita%253ADITA1%255C0dita-1.2-specification.ditamap%253Fv%255C2DITA1.2-spec-XHTML;breadcrumb=urn%253Adita%253ADITA1%255C0architectural_specification.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C3-base.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0ditamarkup.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0dita%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML-urn%253Adita%253ADITA1%255C0archSpec%255C0reconciling%255C3-topic%255C3-and%255C3-map%255C3-metadata.dita%253Fv%255C2DITA1.2%255C3-spec%255C3-XHTML), othermeta and source metadata (among others) should not cascade to children. This bug has been introduced has part of the fix for #1086. The fix may be to revert the incriminated commit (the one that I am unable to find).
process
othermeta and source metadata should not cascade to child topicref here is an example of the present behaviour the map xml simple example the topic xml the child should not have any meta the result xhtml transformation of the topic xml doctype html public dtd xhtml transitional en the child the child should not have any meta as you can see the othermeta has been inherited by the topic but as stated in the othermeta and source metadata among others should not cascade to children this bug has been introduced has part of the fix for the fix may be to revert the incriminated commit the one that i am unable to find
1
1,322
3,874,100,994
IssuesEvent
2016-04-11 19:18:14
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
opened
Remove interventions named "none"
bug Processors
http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374 has a intervention named "none" ```javascript { "brief_summary": "", "id": "13181266-923d-4d28-8e65-73d174314374", "interventions": [ { "attributes": { "id": "94cca51c-9253-4e78-a0ed-766c21e3d083", "name": "none" } } ], "locations": [ { "attributes": { "id": "c2905aaf-49fa-40cb-b858-1dc2e3b484f7", "name": "Japan", "type": "country" }, "role": "recruitment_countries" } ], "organisations": [], "persons": [], "problems": [ { "attributes": { "id": "3e09105d-a4e0-422a-b1a3-2c43c8f0ed60", "name": "Diabetes" } } ], "public_title": "Investigation of the progression of atherosclerosis and atherosclerotic risk factor in patients with diabetes", "registration_date": "2016-04-09T00:00:00.000Z", "url": "http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374" } ```
1.0
Remove interventions named "none" - http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374 has a intervention named "none" ```javascript { "brief_summary": "", "id": "13181266-923d-4d28-8e65-73d174314374", "interventions": [ { "attributes": { "id": "94cca51c-9253-4e78-a0ed-766c21e3d083", "name": "none" } } ], "locations": [ { "attributes": { "id": "c2905aaf-49fa-40cb-b858-1dc2e3b484f7", "name": "Japan", "type": "country" }, "role": "recruitment_countries" } ], "organisations": [], "persons": [], "problems": [ { "attributes": { "id": "3e09105d-a4e0-422a-b1a3-2c43c8f0ed60", "name": "Diabetes" } } ], "public_title": "Investigation of the progression of atherosclerosis and atherosclerotic risk factor in patients with diabetes", "registration_date": "2016-04-09T00:00:00.000Z", "url": "http://api.opentrials.net/v1/trials/13181266-923d-4d28-8e65-73d174314374" } ```
process
remove interventions named none has a intervention named none javascript brief summary id interventions attributes id name none locations attributes id name japan type country role recruitment countries organisations persons problems attributes id name diabetes public title investigation of the progression of atherosclerosis and atherosclerotic risk factor in patients with diabetes registration date url
1
12,186
14,742,224,770
IssuesEvent
2021-01-07 11:55:26
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Keener - Recap of concerns and suggestions - Manual Invoices Billing Cycles
anc-process anp-1 ant-enhancement
In GitLab by @kdjstudios on Mar 27, 2019, 14:16 **Submitted by:** Gaylan **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/7808504 **Server:** All **Client/Site:** All **Account:** All **Issue:** I just wanted to send out a recap of our phone conversation to you and our support team. First, we have a two concerns with how manual invoices are being handled. For manual invoices that are being created between billing cycles we would like to have the ability to change the dates to help clients better distinguish between the manual invoice and the billing cycle invoice. Also this should then help with not having the multiple invoices show on the Invoice Journal Report as the invoices could have different invoice dates. The second part of this concern is when processing a billing cycle both the manual invoice and the billing cycle invoices are being sent at the send step. We would like to have the ability to only send the Billing cycle invoice. As clients are confused when they receive two invoices on the same date. I will provide this feedback to the support team and follow up with their responses. Second, we are wondering if there is a report that can be run for any given date and would show us how many active accounts there were on that date. I thought the Account details report would do this, but it only runs for the current date and is unable to filter to only active accounts. I will look further into this and see if there are any reports that would accomplish this. Thank you again for the feedback and please let us know if there are any other concerns or questions you have.
1.0
Keener - Recap of concerns and suggestions - Manual Invoices Billing Cycles - In GitLab by @kdjstudios on Mar 27, 2019, 14:16 **Submitted by:** Gaylan **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/7808504 **Server:** All **Client/Site:** All **Account:** All **Issue:** I just wanted to send out a recap of our phone conversation to you and our support team. First, we have a two concerns with how manual invoices are being handled. For manual invoices that are being created between billing cycles we would like to have the ability to change the dates to help clients better distinguish between the manual invoice and the billing cycle invoice. Also this should then help with not having the multiple invoices show on the Invoice Journal Report as the invoices could have different invoice dates. The second part of this concern is when processing a billing cycle both the manual invoice and the billing cycle invoices are being sent at the send step. We would like to have the ability to only send the Billing cycle invoice. As clients are confused when they receive two invoices on the same date. I will provide this feedback to the support team and follow up with their responses. Second, we are wondering if there is a report that can be run for any given date and would show us how many active accounts there were on that date. I thought the Account details report would do this, but it only runs for the current date and is unable to filter to only active accounts. I will look further into this and see if there are any reports that would accomplish this. Thank you again for the feedback and please let us know if there are any other concerns or questions you have.
process
keener recap of concerns and suggestions manual invoices billing cycles in gitlab by kdjstudios on mar submitted by gaylan helpdesk server all client site all account all issue i just wanted to send out a recap of our phone conversation to you and our support team first we have a two concerns with how manual invoices are being handled for manual invoices that are being created between billing cycles we would like to have the ability to change the dates to help clients better distinguish between the manual invoice and the billing cycle invoice also this should then help with not having the multiple invoices show on the invoice journal report as the invoices could have different invoice dates the second part of this concern is when processing a billing cycle both the manual invoice and the billing cycle invoices are being sent at the send step we would like to have the ability to only send the billing cycle invoice as clients are confused when they receive two invoices on the same date i will provide this feedback to the support team and follow up with their responses second we are wondering if there is a report that can be run for any given date and would show us how many active accounts there were on that date i thought the account details report would do this but it only runs for the current date and is unable to filter to only active accounts i will look further into this and see if there are any reports that would accomplish this thank you again for the feedback and please let us know if there are any other concerns or questions you have
1
6,226
9,167,379,229
IssuesEvent
2019-03-02 13:09:54
googleapis/google-cloud-cpp
https://api.github.com/repos/googleapis/google-cloud-cpp
closed
Cleanup Kokoro + Microsoft builds.
type: cleanup type: process
Now that `vcpkg` has the same version of `googletest` that we use we can simplify the `ci/kokoro/*.ps1` scripts. No need to use an external project.
1.0
Cleanup Kokoro + Microsoft builds. - Now that `vcpkg` has the same version of `googletest` that we use we can simplify the `ci/kokoro/*.ps1` scripts. No need to use an external project.
process
cleanup kokoro microsoft builds now that vcpkg has the same version of googletest that we use we can simplify the ci kokoro scripts no need to use an external project
1
19,628
25,986,178,516
IssuesEvent
2022-12-20 00:34:38
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
Analista de Negócios de TI na [Grupo LM]
SALVADOR REQUISITOS MODELAGEM DE PROCESSOS HELP WANTED NEGÓCIOS Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Local - Salvador ## Benefícios - Plano de saúde - Transporte - Assistência Odontológica - Ticket Refeição ## Requisitos **Obrigatórios:** - Formação superior completa - Conhecimento em análise com foco no usuário, orientação ao negócio e visão de todo - Experiência com levantamento de requisitos e modelagem de processos - Experiência anterior na área ## Nossa empresa - Com 45 anos de atividades, e atuação nacional, o Grupo LM (www.grupolm.com.br) é constituído pelas empresas LM Frotas, que atua no segmento de terceirização de frotas corporativas, pela concessionária Bravo Caminhões e Ônibus e também pela AuraBrasil, que opera na área de locação de máquinas e equipamentos. ## Como se candidatar - Envie seu curriculo para marina.santos@grupolm.com.br com assunto "Analista de Negócios"
1.0
Analista de Negócios de TI na [Grupo LM] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Local - Salvador ## Benefícios - Plano de saúde - Transporte - Assistência Odontológica - Ticket Refeição ## Requisitos **Obrigatórios:** - Formação superior completa - Conhecimento em análise com foco no usuário, orientação ao negócio e visão de todo - Experiência com levantamento de requisitos e modelagem de processos - Experiência anterior na área ## Nossa empresa - Com 45 anos de atividades, e atuação nacional, o Grupo LM (www.grupolm.com.br) é constituído pelas empresas LM Frotas, que atua no segmento de terceirização de frotas corporativas, pela concessionária Bravo Caminhões e Ônibus e também pela AuraBrasil, que opera na área de locação de máquinas e equipamentos. ## Como se candidatar - Envie seu curriculo para marina.santos@grupolm.com.br com assunto "Analista de Negócios"
process
analista de negócios de ti na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na local salvador benefícios plano de saúde transporte assistência odontológica ticket refeição requisitos obrigatórios formação superior completa conhecimento em análise com foco no usuário orientação ao negócio e visão de todo experiência com levantamento de requisitos e modelagem de processos experiência anterior na área nossa empresa com anos de atividades e atuação nacional o grupo lm é constituído pelas empresas lm frotas que atua no segmento de terceirização de frotas corporativas pela concessionária bravo caminhões e ônibus e também pela aurabrasil que opera na área de locação de máquinas e equipamentos como se candidatar envie seu curriculo para marina santos grupolm com br com assunto analista de negócios
1
10,406
13,204,215,990
IssuesEvent
2020-08-14 15:30:09
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
The difference between `pipeline` and `source` is not clear.
Pri2 devops-cicd-process/tech devops/prod doc-enhancement
The difference between `pipeline` and `source` is not clear. In the example they seem to be the same apart from small spelling differences leaving me wonder what both of their meaning truly is. Can they both be the same? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 86285f72-9e28-da97-59bb-c29eb60f627d * Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54 * Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @ashkir * Microsoft Alias: **ashkir**
1.0
The difference between `pipeline` and `source` is not clear. - The difference between `pipeline` and `source` is not clear. In the example they seem to be the same apart from small spelling differences leaving me wonder what both of their meaning truly is. Can they both be the same? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 86285f72-9e28-da97-59bb-c29eb60f627d * Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54 * Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @ashkir * Microsoft Alias: **ashkir**
process
the difference between pipeline and source is not clear the difference between pipeline and source is not clear in the example they seem to be the same apart from small spelling differences leaving me wonder what both of their meaning truly is can they both be the same document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login ashkir microsoft alias ashkir
1
5,101
7,881,686,895
IssuesEvent
2018-06-26 19:56:29
ArctosDB/new-collections
https://api.github.com/repos/ArctosDB/new-collections
closed
Angelo State Collection Mentor(s)
Application in process
Assign collection mentor Default = AWG Chair or Vice-Chair AWG member can volunteer to act as primary contact, especially if they have similar collections or specific knowledge about a collection; can serve as ‘in kind support’ for collections to help offset costs
1.0
Angelo State Collection Mentor(s) - Assign collection mentor Default = AWG Chair or Vice-Chair AWG member can volunteer to act as primary contact, especially if they have similar collections or specific knowledge about a collection; can serve as ‘in kind support’ for collections to help offset costs
process
angelo state collection mentor s assign collection mentor default awg chair or vice chair awg member can volunteer to act as primary contact especially if they have similar collections or specific knowledge about a collection can serve as ‘in kind support’ for collections to help offset costs
1
21,488
29,578,222,797
IssuesEvent
2023-06-07 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Wed, 7 Jun 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Real-Time Onboard Object Detection for Augmented Reality: Enhancing Head-Mounted Display with YOLOv8 - **Authors:** Mikołaj Łysakowski, Kamil Żywanowski, Adam Banaszczyk, Michał R. Nowicki, Piotr Skrzypczyński, Sławomir K. Tadeja - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Human-Computer Interaction (cs.HC) - **Arxiv link:** https://arxiv.org/abs/2306.03537 - **Pdf link:** https://arxiv.org/pdf/2306.03537 - **Abstract** This paper introduces a software architecture for real-time object detection using machine learning (ML) in an augmented reality (AR) environment. Our approach uses the recent state-of-the-art YOLOv8 network that runs onboard on the Microsoft HoloLens 2 head-mounted display (HMD). The primary motivation behind this research is to enable the application of advanced ML models for enhanced perception and situational awareness with a wearable, hands-free AR platform. We show the image processing pipeline for the YOLOv8 model and the techniques used to make it real-time on the resource-limited edge computing platform of the headset. The experimental results demonstrate that our solution achieves real-time processing without needing offloading tasks to the cloud or any other external servers while retaining satisfactory accuracy regarding the usual mAP metric and measured qualitative performance ### SDR-GAIN: A High Real-Time Occluded Pedestrian Pose Completion Method for Autonomous Driving - **Authors:** Honghao Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2306.03538 - **Pdf link:** https://arxiv.org/pdf/2306.03538 - **Abstract** To mitigate the challenges arising from partial occlusion in human pose keypoint based pedestrian detection methods , we present a novel pedestrian pose keypoint completion method called the separation and dimensionality reduction-based generative adversarial imputation networks (SDR-GAIN) . Firstly, we utilize OpenPose to estimate pedestrian poses in images. Then, we isolate the head and torso keypoints of pedestrians with incomplete keypoints due to occlusion or other factors and perform dimensionality reduction to enhance features and further unify feature distribution. Finally, we introduce two generative models based on the generative adversarial networks (GAN) framework, which incorporate Huber loss, residual structure, and L1 regularization to generate missing parts of the incomplete head and torso pose keypoints of partially occluded pedestrians, resulting in pose completion. Our experiments on MS COCO and JAAD datasets demonstrate that SDR-GAIN outperforms basic GAIN framework, interpolation methods PCHIP and MAkima, machine learning methods k-NN and MissForest in terms of pose completion task. In addition, the runtime of SDR-GAIN is approximately 0.4ms, displaying high real-time performance and significant application value in the field of autonomous driving. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### MoDAR: Using Motion Forecasting for 3D Object Detection in Point Cloud Sequences - **Authors:** Yingwei Li, Charles R. Qi, Yin Zhou, Chenxi Liu, Dragomir Anguelov - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.03206 - **Pdf link:** https://arxiv.org/pdf/2306.03206 - **Abstract** Occluded and long-range objects are ubiquitous and challenging for 3D object detection. Point cloud sequence data provide unique opportunities to improve such cases, as an occluded or distant object can be observed from different viewpoints or gets better visibility over time. However, the efficiency and effectiveness in encoding long-term sequence data can still be improved. In this work, we propose MoDAR, using motion forecasting outputs as a type of virtual modality, to augment LiDAR point clouds. The MoDAR modality propagates object information from temporal contexts to a target frame, represented as a set of virtual points, one for each object from a waypoint on a forecasted trajectory. A fused point cloud of both raw sensor points and the virtual points can then be fed to any off-the-shelf point-cloud based 3D object detector. Evaluated on the Waymo Open Dataset, our method significantly improves prior art detectors by using motion forecasting from extra-long sequences (e.g. 18 seconds), achieving new state of the arts, while not adding much computation overhead. ### MetaGait: Learning to Learn an Omni Sample Adaptive Representation for Gait Recognition - **Authors:** Huanzhang Dou, Pengyi Zhang, Wei Su, Yunlong Yu, Xi Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2306.03445 - **Pdf link:** https://arxiv.org/pdf/2306.03445 - **Abstract** Gait recognition, which aims at identifying individuals by their walking patterns, has recently drawn increasing research attention. However, gait recognition still suffers from the conflicts between the limited binary visual clues of the silhouette and numerous covariates with diverse scales, which brings challenges to the model's adaptiveness. In this paper, we address this conflict by developing a novel MetaGait that learns to learn an omni sample adaptive representation. Towards this goal, MetaGait injects meta-knowledge, which could guide the model to perceive sample-specific properties, into the calibration network of the attention mechanism to improve the adaptiveness from the omni-scale, omni-dimension, and omni-process perspectives. Specifically, we leverage the meta-knowledge across the entire process, where Meta Triple Attention and Meta Temporal Pooling are presented respectively to adaptively capture omni-scale dependency from spatial/channel/temporal dimensions simultaneously and to adaptively aggregate temporal information through integrating the merits of three complementary temporal aggregation methods. Extensive experiments demonstrate the state-of-the-art performance of the proposed MetaGait. On CASIA-B, we achieve rank-1 accuracy of 98.7%, 96.0%, and 89.3% under three conditions, respectively. On OU-MVLP, we achieve rank-1 accuracy of 92.4%. ### Instructive Feature Enhancement for Dichotomous Medical Image Segmentation - **Authors:** Lian Liu, Han Zhou, Jiongquan Chen, Sijing Liu, Wenlong Shi, Dong Ni, Deng-Ping Fan, Xin Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.03497 - **Pdf link:** https://arxiv.org/pdf/2306.03497 - **Abstract** Deep neural networks have been widely applied in dichotomous medical image segmentation (DMIS) of many anatomical structures in several modalities, achieving promising performance. However, existing networks tend to struggle with task-specific, heavy and complex designs to improve accuracy. They made little instructions to which feature channels would be more beneficial for segmentation, and that may be why the performance and universality of these segmentation models are hindered. In this study, we propose an instructive feature enhancement approach, namely IFE, to adaptively select feature channels with rich texture cues and strong discriminability to enhance raw features based on local curvature or global information entropy criteria. Being plug-and-play and applicable for diverse DMIS tasks, IFE encourages the model to focus on texture-rich features which are especially important for the ambiguous and challenging boundary identification, simultaneously achieving simplicity, universality, and certain interpretability. To evaluate the proposed IFE, we constructed the first large-scale DMIS dataset Cosmos55k, which contains 55,023 images from 7 modalities and 26 anatomical structures. Extensive experiments show that IFE can improve the performance of classic segmentation networks across different anatomies and modalities with only slight modifications. Code is available at https://github.com/yezi-66/IFE ### RDFC-GAN: RGB-Depth Fusion CycleGAN for Indoor Depth Completion - **Authors:** Haowen Wang, Zhengping Che, Mingyuan Wang, Zhiyuan Xu, Xiuquan Qiao, Mengshi Qi, Feifei Feng, Jian Tang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2306.03584 - **Pdf link:** https://arxiv.org/pdf/2306.03584 - **Abstract** The raw depth image captured by indoor depth sensors usually has an extensive range of missing depth values due to inherent limitations such as the inability to perceive transparent objects and the limited distance range. The incomplete depth map with missing values burdens many downstream vision tasks, and a rising number of depth completion methods have been proposed to alleviate this issue. While most existing methods can generate accurate dense depth maps from sparse and uniformly sampled depth maps, they are not suitable for complementing large contiguous regions of missing depth values, which is common and critical in images captured in indoor environments. To overcome these challenges, we design a novel two-branch end-to-end fusion network named RDFC-GAN, which takes a pair of RGB and incomplete depth images as input to predict a dense and completed depth map. The first branch employs an encoder-decoder structure, by adhering to the Manhattan world assumption and utilizing normal maps from RGB-D information as guidance, to regress the local dense depth values from the raw depth map. In the other branch, we propose an RGB-depth fusion CycleGAN to transfer the RGB image to the fine-grained textured depth map. We adopt adaptive fusion modules named W-AdaIN to propagate the features across the two branches, and we append a confidence fusion head to fuse the two outputs of the branches for the final depth map. Extensive experiments on NYU-Depth V2 and SUN RGB-D demonstrate that our proposed method clearly improves the depth completion performance, especially in a more realistic setting of indoor environments, with the help of our proposed pseudo depth maps in training. ## Keyword: raw image There is no result
2.0
New submissions for Wed, 7 Jun 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Real-Time Onboard Object Detection for Augmented Reality: Enhancing Head-Mounted Display with YOLOv8 - **Authors:** Mikołaj Łysakowski, Kamil Żywanowski, Adam Banaszczyk, Michał R. Nowicki, Piotr Skrzypczyński, Sławomir K. Tadeja - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Human-Computer Interaction (cs.HC) - **Arxiv link:** https://arxiv.org/abs/2306.03537 - **Pdf link:** https://arxiv.org/pdf/2306.03537 - **Abstract** This paper introduces a software architecture for real-time object detection using machine learning (ML) in an augmented reality (AR) environment. Our approach uses the recent state-of-the-art YOLOv8 network that runs onboard on the Microsoft HoloLens 2 head-mounted display (HMD). The primary motivation behind this research is to enable the application of advanced ML models for enhanced perception and situational awareness with a wearable, hands-free AR platform. We show the image processing pipeline for the YOLOv8 model and the techniques used to make it real-time on the resource-limited edge computing platform of the headset. The experimental results demonstrate that our solution achieves real-time processing without needing offloading tasks to the cloud or any other external servers while retaining satisfactory accuracy regarding the usual mAP metric and measured qualitative performance ### SDR-GAIN: A High Real-Time Occluded Pedestrian Pose Completion Method for Autonomous Driving - **Authors:** Honghao Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2306.03538 - **Pdf link:** https://arxiv.org/pdf/2306.03538 - **Abstract** To mitigate the challenges arising from partial occlusion in human pose keypoint based pedestrian detection methods , we present a novel pedestrian pose keypoint completion method called the separation and dimensionality reduction-based generative adversarial imputation networks (SDR-GAIN) . Firstly, we utilize OpenPose to estimate pedestrian poses in images. Then, we isolate the head and torso keypoints of pedestrians with incomplete keypoints due to occlusion or other factors and perform dimensionality reduction to enhance features and further unify feature distribution. Finally, we introduce two generative models based on the generative adversarial networks (GAN) framework, which incorporate Huber loss, residual structure, and L1 regularization to generate missing parts of the incomplete head and torso pose keypoints of partially occluded pedestrians, resulting in pose completion. Our experiments on MS COCO and JAAD datasets demonstrate that SDR-GAIN outperforms basic GAIN framework, interpolation methods PCHIP and MAkima, machine learning methods k-NN and MissForest in terms of pose completion task. In addition, the runtime of SDR-GAIN is approximately 0.4ms, displaying high real-time performance and significant application value in the field of autonomous driving. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### MoDAR: Using Motion Forecasting for 3D Object Detection in Point Cloud Sequences - **Authors:** Yingwei Li, Charles R. Qi, Yin Zhou, Chenxi Liu, Dragomir Anguelov - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.03206 - **Pdf link:** https://arxiv.org/pdf/2306.03206 - **Abstract** Occluded and long-range objects are ubiquitous and challenging for 3D object detection. Point cloud sequence data provide unique opportunities to improve such cases, as an occluded or distant object can be observed from different viewpoints or gets better visibility over time. However, the efficiency and effectiveness in encoding long-term sequence data can still be improved. In this work, we propose MoDAR, using motion forecasting outputs as a type of virtual modality, to augment LiDAR point clouds. The MoDAR modality propagates object information from temporal contexts to a target frame, represented as a set of virtual points, one for each object from a waypoint on a forecasted trajectory. A fused point cloud of both raw sensor points and the virtual points can then be fed to any off-the-shelf point-cloud based 3D object detector. Evaluated on the Waymo Open Dataset, our method significantly improves prior art detectors by using motion forecasting from extra-long sequences (e.g. 18 seconds), achieving new state of the arts, while not adding much computation overhead. ### MetaGait: Learning to Learn an Omni Sample Adaptive Representation for Gait Recognition - **Authors:** Huanzhang Dou, Pengyi Zhang, Wei Su, Yunlong Yu, Xi Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2306.03445 - **Pdf link:** https://arxiv.org/pdf/2306.03445 - **Abstract** Gait recognition, which aims at identifying individuals by their walking patterns, has recently drawn increasing research attention. However, gait recognition still suffers from the conflicts between the limited binary visual clues of the silhouette and numerous covariates with diverse scales, which brings challenges to the model's adaptiveness. In this paper, we address this conflict by developing a novel MetaGait that learns to learn an omni sample adaptive representation. Towards this goal, MetaGait injects meta-knowledge, which could guide the model to perceive sample-specific properties, into the calibration network of the attention mechanism to improve the adaptiveness from the omni-scale, omni-dimension, and omni-process perspectives. Specifically, we leverage the meta-knowledge across the entire process, where Meta Triple Attention and Meta Temporal Pooling are presented respectively to adaptively capture omni-scale dependency from spatial/channel/temporal dimensions simultaneously and to adaptively aggregate temporal information through integrating the merits of three complementary temporal aggregation methods. Extensive experiments demonstrate the state-of-the-art performance of the proposed MetaGait. On CASIA-B, we achieve rank-1 accuracy of 98.7%, 96.0%, and 89.3% under three conditions, respectively. On OU-MVLP, we achieve rank-1 accuracy of 92.4%. ### Instructive Feature Enhancement for Dichotomous Medical Image Segmentation - **Authors:** Lian Liu, Han Zhou, Jiongquan Chen, Sijing Liu, Wenlong Shi, Dong Ni, Deng-Ping Fan, Xin Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.03497 - **Pdf link:** https://arxiv.org/pdf/2306.03497 - **Abstract** Deep neural networks have been widely applied in dichotomous medical image segmentation (DMIS) of many anatomical structures in several modalities, achieving promising performance. However, existing networks tend to struggle with task-specific, heavy and complex designs to improve accuracy. They made little instructions to which feature channels would be more beneficial for segmentation, and that may be why the performance and universality of these segmentation models are hindered. In this study, we propose an instructive feature enhancement approach, namely IFE, to adaptively select feature channels with rich texture cues and strong discriminability to enhance raw features based on local curvature or global information entropy criteria. Being plug-and-play and applicable for diverse DMIS tasks, IFE encourages the model to focus on texture-rich features which are especially important for the ambiguous and challenging boundary identification, simultaneously achieving simplicity, universality, and certain interpretability. To evaluate the proposed IFE, we constructed the first large-scale DMIS dataset Cosmos55k, which contains 55,023 images from 7 modalities and 26 anatomical structures. Extensive experiments show that IFE can improve the performance of classic segmentation networks across different anatomies and modalities with only slight modifications. Code is available at https://github.com/yezi-66/IFE ### RDFC-GAN: RGB-Depth Fusion CycleGAN for Indoor Depth Completion - **Authors:** Haowen Wang, Zhengping Che, Mingyuan Wang, Zhiyuan Xu, Xiuquan Qiao, Mengshi Qi, Feifei Feng, Jian Tang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2306.03584 - **Pdf link:** https://arxiv.org/pdf/2306.03584 - **Abstract** The raw depth image captured by indoor depth sensors usually has an extensive range of missing depth values due to inherent limitations such as the inability to perceive transparent objects and the limited distance range. The incomplete depth map with missing values burdens many downstream vision tasks, and a rising number of depth completion methods have been proposed to alleviate this issue. While most existing methods can generate accurate dense depth maps from sparse and uniformly sampled depth maps, they are not suitable for complementing large contiguous regions of missing depth values, which is common and critical in images captured in indoor environments. To overcome these challenges, we design a novel two-branch end-to-end fusion network named RDFC-GAN, which takes a pair of RGB and incomplete depth images as input to predict a dense and completed depth map. The first branch employs an encoder-decoder structure, by adhering to the Manhattan world assumption and utilizing normal maps from RGB-D information as guidance, to regress the local dense depth values from the raw depth map. In the other branch, we propose an RGB-depth fusion CycleGAN to transfer the RGB image to the fine-grained textured depth map. We adopt adaptive fusion modules named W-AdaIN to propagate the features across the two branches, and we append a confidence fusion head to fuse the two outputs of the branches for the final depth map. Extensive experiments on NYU-Depth V2 and SUN RGB-D demonstrate that our proposed method clearly improves the depth completion performance, especially in a more realistic setting of indoor environments, with the help of our proposed pseudo depth maps in training. ## Keyword: raw image There is no result
process
new submissions for wed jun keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp real time onboard object detection for augmented reality enhancing head mounted display with authors mikołaj łysakowski kamil żywanowski adam banaszczyk michał r nowicki piotr skrzypczyński sławomir k tadeja subjects computer vision and pattern recognition cs cv human computer interaction cs hc arxiv link pdf link abstract this paper introduces a software architecture for real time object detection using machine learning ml in an augmented reality ar environment our approach uses the recent state of the art network that runs onboard on the microsoft hololens head mounted display hmd the primary motivation behind this research is to enable the application of advanced ml models for enhanced perception and situational awareness with a wearable hands free ar platform we show the image processing pipeline for the model and the techniques used to make it real time on the resource limited edge computing platform of the headset the experimental results demonstrate that our solution achieves real time processing without needing offloading tasks to the cloud or any other external servers while retaining satisfactory accuracy regarding the usual map metric and measured qualitative performance sdr gain a high real time occluded pedestrian pose completion method for autonomous driving authors honghao fu subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract to mitigate the challenges arising from partial occlusion in human pose keypoint based pedestrian detection methods we present a novel pedestrian pose keypoint completion method called the separation and dimensionality reduction based generative adversarial imputation networks sdr gain firstly we utilize openpose to estimate pedestrian poses in images then we isolate the head and torso keypoints of pedestrians with incomplete keypoints due to occlusion or other factors and perform dimensionality reduction to enhance features and further unify feature distribution finally we introduce two generative models based on the generative adversarial networks gan framework which incorporate huber loss residual structure and regularization to generate missing parts of the incomplete head and torso pose keypoints of partially occluded pedestrians resulting in pose completion our experiments on ms coco and jaad datasets demonstrate that sdr gain outperforms basic gain framework interpolation methods pchip and makima machine learning methods k nn and missforest in terms of pose completion task in addition the runtime of sdr gain is approximately displaying high real time performance and significant application value in the field of autonomous driving keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw modar using motion forecasting for object detection in point cloud sequences authors yingwei li charles r qi yin zhou chenxi liu dragomir anguelov subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract occluded and long range objects are ubiquitous and challenging for object detection point cloud sequence data provide unique opportunities to improve such cases as an occluded or distant object can be observed from different viewpoints or gets better visibility over time however the efficiency and effectiveness in encoding long term sequence data can still be improved in this work we propose modar using motion forecasting outputs as a type of virtual modality to augment lidar point clouds the modar modality propagates object information from temporal contexts to a target frame represented as a set of virtual points one for each object from a waypoint on a forecasted trajectory a fused point cloud of both raw sensor points and the virtual points can then be fed to any off the shelf point cloud based object detector evaluated on the waymo open dataset our method significantly improves prior art detectors by using motion forecasting from extra long sequences e g seconds achieving new state of the arts while not adding much computation overhead metagait learning to learn an omni sample adaptive representation for gait recognition authors huanzhang dou pengyi zhang wei su yunlong yu xi li subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract gait recognition which aims at identifying individuals by their walking patterns has recently drawn increasing research attention however gait recognition still suffers from the conflicts between the limited binary visual clues of the silhouette and numerous covariates with diverse scales which brings challenges to the model s adaptiveness in this paper we address this conflict by developing a novel metagait that learns to learn an omni sample adaptive representation towards this goal metagait injects meta knowledge which could guide the model to perceive sample specific properties into the calibration network of the attention mechanism to improve the adaptiveness from the omni scale omni dimension and omni process perspectives specifically we leverage the meta knowledge across the entire process where meta triple attention and meta temporal pooling are presented respectively to adaptively capture omni scale dependency from spatial channel temporal dimensions simultaneously and to adaptively aggregate temporal information through integrating the merits of three complementary temporal aggregation methods extensive experiments demonstrate the state of the art performance of the proposed metagait on casia b we achieve rank accuracy of and under three conditions respectively on ou mvlp we achieve rank accuracy of instructive feature enhancement for dichotomous medical image segmentation authors lian liu han zhou jiongquan chen sijing liu wenlong shi dong ni deng ping fan xin yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep neural networks have been widely applied in dichotomous medical image segmentation dmis of many anatomical structures in several modalities achieving promising performance however existing networks tend to struggle with task specific heavy and complex designs to improve accuracy they made little instructions to which feature channels would be more beneficial for segmentation and that may be why the performance and universality of these segmentation models are hindered in this study we propose an instructive feature enhancement approach namely ife to adaptively select feature channels with rich texture cues and strong discriminability to enhance raw features based on local curvature or global information entropy criteria being plug and play and applicable for diverse dmis tasks ife encourages the model to focus on texture rich features which are especially important for the ambiguous and challenging boundary identification simultaneously achieving simplicity universality and certain interpretability to evaluate the proposed ife we constructed the first large scale dmis dataset which contains images from modalities and anatomical structures extensive experiments show that ife can improve the performance of classic segmentation networks across different anatomies and modalities with only slight modifications code is available at rdfc gan rgb depth fusion cyclegan for indoor depth completion authors haowen wang zhengping che mingyuan wang zhiyuan xu xiuquan qiao mengshi qi feifei feng jian tang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract the raw depth image captured by indoor depth sensors usually has an extensive range of missing depth values due to inherent limitations such as the inability to perceive transparent objects and the limited distance range the incomplete depth map with missing values burdens many downstream vision tasks and a rising number of depth completion methods have been proposed to alleviate this issue while most existing methods can generate accurate dense depth maps from sparse and uniformly sampled depth maps they are not suitable for complementing large contiguous regions of missing depth values which is common and critical in images captured in indoor environments to overcome these challenges we design a novel two branch end to end fusion network named rdfc gan which takes a pair of rgb and incomplete depth images as input to predict a dense and completed depth map the first branch employs an encoder decoder structure by adhering to the manhattan world assumption and utilizing normal maps from rgb d information as guidance to regress the local dense depth values from the raw depth map in the other branch we propose an rgb depth fusion cyclegan to transfer the rgb image to the fine grained textured depth map we adopt adaptive fusion modules named w adain to propagate the features across the two branches and we append a confidence fusion head to fuse the two outputs of the branches for the final depth map extensive experiments on nyu depth and sun rgb d demonstrate that our proposed method clearly improves the depth completion performance especially in a more realistic setting of indoor environments with the help of our proposed pseudo depth maps in training keyword raw image there is no result
1
817,337
30,637,646,630
IssuesEvent
2023-07-24 19:05:15
priyankarpal/ProjectsHut
https://api.github.com/repos/priyankarpal/ProjectsHut
closed
docs: Adding a MIT Licenses badge and some more text to Licenses section in readme
📄 aspect: text good first issue ✨ goal: improvement 🟨 priority: medium 🚦status: awaiting triage
### what's wrong with the documentation? The readme file of the repo should have a MIT Licenses badge and some more text regarding licenses in the Licenses section to make it more professional. ### Add screenshots ![afters](https://github.com/Eduhub-Community/Eduhub-Community.github.io/assets/116020663/df57e4ed-2b1c-4236-b95d-17c9f3190e5a) ### Record - [X] I have checked the existing [issues](https://github.com/priyankarpal/ProjectsHut/issues) - [X] I have read the [Contributing Guidelines](https://github.com/priyankarpal/ProjectsHut/blob/main/contributing.md) - [X] I agree to follow this project's [Code of Conduct](https://github.com/priyankarpal/ProjectsHut/blob/main/CODE_OF_CONDUCT.md) - [X] I'm a GSSoC'23 contributor - [X] I want to work on this issue
1.0
docs: Adding a MIT Licenses badge and some more text to Licenses section in readme - ### what's wrong with the documentation? The readme file of the repo should have a MIT Licenses badge and some more text regarding licenses in the Licenses section to make it more professional. ### Add screenshots ![afters](https://github.com/Eduhub-Community/Eduhub-Community.github.io/assets/116020663/df57e4ed-2b1c-4236-b95d-17c9f3190e5a) ### Record - [X] I have checked the existing [issues](https://github.com/priyankarpal/ProjectsHut/issues) - [X] I have read the [Contributing Guidelines](https://github.com/priyankarpal/ProjectsHut/blob/main/contributing.md) - [X] I agree to follow this project's [Code of Conduct](https://github.com/priyankarpal/ProjectsHut/blob/main/CODE_OF_CONDUCT.md) - [X] I'm a GSSoC'23 contributor - [X] I want to work on this issue
non_process
docs adding a mit licenses badge and some more text to licenses section in readme what s wrong with the documentation the readme file of the repo should have a mit licenses badge and some more text regarding licenses in the licenses section to make it more professional add screenshots record i have checked the existing i have read the i agree to follow this project s i m a gssoc contributor i want to work on this issue
0
11,243
14,015,299,412
IssuesEvent
2020-10-29 13:09:05
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - dc:language
Process - implement Term - change
## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): The recommended specification has been deprecated. * Proponents (who needs this change): Everyone Proposed new attributes of the term: * Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as RFC 5646.
1.0
Change term - dc:language - ## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): The recommended specification has been deprecated. * Proponents (who needs this change): Everyone Proposed new attributes of the term: * Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as RFC 5646.
process
change term dc language change term submitter john wieczorek justification why is this change necessary the recommended specification has been deprecated proponents who needs this change everyone proposed new attributes of the term usage comments recommendations regarding content etc recommended best practice is to use a controlled vocabulary such as rfc
1
7,939
8,125,087,465
IssuesEvent
2018-08-16 19:44:14
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
AMQP .NET: Broken Anchor Link
service-bus-messaging/svc
Priority: Minor Type: Broken Link Looks like https://github.com/MicrosoftDocs/azure-docs/commit/ed466dcd1b4f3a690d08f04288cb48f98ac37003 is what changed the header, so change line 28 from: ``` In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#unsupported-features-restrictions-and-behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP. ``` to ``` In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP. ``` Potentially reword as well since the heading no longer has "unsupported features" in it. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e28d060c-a765-ed3a-9de5-7ba9ca93d649 * Version Independent ID: e6e933ef-50d7-712e-fb4b-e7ef555fb1b6 * Content: [Azure Service Bus with .NET and AMQP 1.0](https://docs.microsoft.com/en-us/azure/service-bus-messaging/service-bus-amqp-dotnet#unsupported-features-restrictions-and-behavioral-differences) * Content Source: [articles/service-bus-messaging/service-bus-amqp-dotnet.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-bus-messaging/service-bus-amqp-dotnet.md) * Service: **service-bus-messaging** * GitHub Login: @sethmanheim * Microsoft Alias: **sethm**
1.0
AMQP .NET: Broken Anchor Link - Priority: Minor Type: Broken Link Looks like https://github.com/MicrosoftDocs/azure-docs/commit/ed466dcd1b4f3a690d08f04288cb48f98ac37003 is what changed the header, so change line 28 from: ``` In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#unsupported-features-restrictions-and-behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP. ``` to ``` In the current release, there are a few API features that are not supported when using AMQP. These unsupported features are listed later in the section [Unsupported features, restrictions, and behavioral differences](#behavioral-differences). Some of the advanced configuration settings also have a different meaning when using AMQP. ``` Potentially reword as well since the heading no longer has "unsupported features" in it. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e28d060c-a765-ed3a-9de5-7ba9ca93d649 * Version Independent ID: e6e933ef-50d7-712e-fb4b-e7ef555fb1b6 * Content: [Azure Service Bus with .NET and AMQP 1.0](https://docs.microsoft.com/en-us/azure/service-bus-messaging/service-bus-amqp-dotnet#unsupported-features-restrictions-and-behavioral-differences) * Content Source: [articles/service-bus-messaging/service-bus-amqp-dotnet.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-bus-messaging/service-bus-amqp-dotnet.md) * Service: **service-bus-messaging** * GitHub Login: @sethmanheim * Microsoft Alias: **sethm**
non_process
amqp net broken anchor link priority minor type broken link looks like is what changed the header so change line from in the current release there are a few api features that are not supported when using amqp these unsupported features are listed later in the section unsupported features restrictions and behavioral differences some of the advanced configuration settings also have a different meaning when using amqp to in the current release there are a few api features that are not supported when using amqp these unsupported features are listed later in the section behavioral differences some of the advanced configuration settings also have a different meaning when using amqp potentially reword as well since the heading no longer has unsupported features in it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service service bus messaging github login sethmanheim microsoft alias sethm
0
18,593
24,569,693,128
IssuesEvent
2022-10-13 07:39:16
googleapis/nodejs-certificate-manager
https://api.github.com/repos/googleapis/nodejs-certificate-manager
closed
Promote library to 1.0.0
type: bug type: process api: secretmanager
We should promote this library to `v1.0.0` if it is no longer pre-release.
1.0
Promote library to 1.0.0 - We should promote this library to `v1.0.0` if it is no longer pre-release.
process
promote library to we should promote this library to if it is no longer pre release
1
15,200
19,011,011,927
IssuesEvent
2021-11-23 09:19:12
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Move non-native Processing providers into independent plugins (Request in QGIS)
Processing 3.22
### Request for documentation From pull request QGIS/qgis#43792 Author: @alexbruy QGIS version: 3.22 **Move non-native Processing providers into independent plugins** ### PR Description: ## Description Following discussion in the [QEP 226](https://github.com/qgis/QGIS-Enhancement-Proposals/issues/226) separate SAGA, GRASS and OTB providers into independent plugins. This should prevent issues with loading of the Processing plugin in cases when one of these providers is broken. Also it will make easier turning them into 3rd party plugins in the future if we will make such decision. SAGA and GRASS plugins are enabled by default to provide the same UX as before. OTB provider can be enabled in the Plugin Manager if necessary. Enabling/disabling of these three providers now performed via the Plugin Manager like with any other plugin. This is mostly an internal change, the only visible for users part is using Plugin Manager to activate or deactivate SAGA, GRASS and OTB providers. ### Commits tagged with [need-docs] or [FEATURE]
1.0
Move non-native Processing providers into independent plugins (Request in QGIS) - ### Request for documentation From pull request QGIS/qgis#43792 Author: @alexbruy QGIS version: 3.22 **Move non-native Processing providers into independent plugins** ### PR Description: ## Description Following discussion in the [QEP 226](https://github.com/qgis/QGIS-Enhancement-Proposals/issues/226) separate SAGA, GRASS and OTB providers into independent plugins. This should prevent issues with loading of the Processing plugin in cases when one of these providers is broken. Also it will make easier turning them into 3rd party plugins in the future if we will make such decision. SAGA and GRASS plugins are enabled by default to provide the same UX as before. OTB provider can be enabled in the Plugin Manager if necessary. Enabling/disabling of these three providers now performed via the Plugin Manager like with any other plugin. This is mostly an internal change, the only visible for users part is using Plugin Manager to activate or deactivate SAGA, GRASS and OTB providers. ### Commits tagged with [need-docs] or [FEATURE]
process
move non native processing providers into independent plugins request in qgis request for documentation from pull request qgis qgis author alexbruy qgis version move non native processing providers into independent plugins pr description description following discussion in the separate saga grass and otb providers into independent plugins this should prevent issues with loading of the processing plugin in cases when one of these providers is broken also it will make easier turning them into party plugins in the future if we will make such decision saga and grass plugins are enabled by default to provide the same ux as before otb provider can be enabled in the plugin manager if necessary enabling disabling of these three providers now performed via the plugin manager like with any other plugin this is mostly an internal change the only visible for users part is using plugin manager to activate or deactivate saga grass and otb providers commits tagged with or
1
293,025
8,971,918,537
IssuesEvent
2019-01-29 16:58:09
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.huffingtonpost.kr - desktop site instead of mobile site
browser-firefox-mobile priority-normal
<!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: Yes **Problem type**: Desktop site instead of mobile site **Description**: font does not optimize as mobile site **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/1/e678f4cf-538b-4c66-982e-12862739276e-thumb.jpeg)](https://webcompat.com/uploads/2019/1/e678f4cf-538b-4c66-982e-12862739276e.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190121125354</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Error: "TypeError: window.YAHOO is undefined" {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 314}]\n@https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver:314:3\n', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/ko_KR/sdk.js#xfbml=1&version=v2.10&appId=284021941754241 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-WN49HC was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=50_50 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@TopRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 370}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@x28 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 430}]', u'[JavaScript Warning: "https://imasdk.googleapis.com/js/sdkloader/ima3.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1366}]', u'[JavaScript Warning: "https://ads.admaru.com/js/ads_huffingtonpost_article_body_us.js?t=1548108177754 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1380}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ads.admaru.com/js/huffingtonpost_admaru_incontent_bigbanner.js?t=1548108177791 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1395}]', u'[console.log([Sticky Ad] > Configuration: {"stickyEnabled":true,"thresholdHeight":109,"stickyDuration":5000,"waitForAdInterval":1500,"waitForAdRetries":150,"offsetFromTop":250}) https://amp.assets.huffpost.com/assets-a380132db9db2f1d3548a6c503316cc8/js/scripts.min.js:2:2905]', u'[JavaScript Warning: "The resource at https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/en_US/fbevents.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=40_40 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.youtube.com/iframe_api was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@BottomRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1815}]', u'[JavaScript Warning: "https://d1z2jf7jlzjs58.cloudfront.net/p.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://connect.facebook.net/en_US/fbevents.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.huffingtonpost.kr - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.0.0; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android 8.0.0 **Tested Another Browser**: Yes **Problem type**: Desktop site instead of mobile site **Description**: font does not optimize as mobile site **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/1/e678f4cf-538b-4c66-982e-12862739276e-thumb.jpeg)](https://webcompat.com/uploads/2019/1/e678f4cf-538b-4c66-982e-12862739276e.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190121125354</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://s.yimg.com/ss/rapid-3.50.5.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Error: "TypeError: window.YAHOO is undefined" {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 314}]\n@https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver:314:3\n', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.twitter.com/widgets.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/ko_KR/sdk.js#xfbml=1&version=v2.10&appId=284021941754241 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.googletagmanager.com/gtm.js?id=GTM-WN49HC was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c456327250000be00c59114.jpeg?cache=wasdhmYxPd&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c419441360000f103680b83.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c401ef223000030001f9e54.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4563fc250000e000c59115.jpeg?cache=6PdFRz5WbS&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45603a360000e204680c16.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565193b00009e016890f6.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565643b00009e016890f7.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566512500003a017db67f.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4566a825000031007db680.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4565c02500003a017db67e.jpeg?ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3fd68a2400009d009f9f86.jpeg?cache=aA9y6fS9Zi&ops=scalefit_630_noupscale was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4172ba2500003600c59067.jpeg?cache=FUNioDhUoU&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3ed0ae250000a2007db4d2.jpeg?cache=ZVu29LLSbb&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bc05f2b1f0000000225950d.jpeg?cache=RdzxEUfyQI&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c3459572200003909dfa223.jpeg?cache=TzKC0Wfezx&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5bade073260000e900810f90.jpeg?cache=ZDc8p1aUZV&ops=250_140 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=50_50 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@TopRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 370}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@x28 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 430}]', u'[JavaScript Warning: "https://imasdk.googleapis.com/js/sdkloader/ima3.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1366}]', u'[JavaScript Warning: "https://ads.admaru.com/js/ads_huffingtonpost_article_body_us.js?t=1548108177754 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1380}]', u'[JavaScript Warning: "The resource at https://www.googletagservices.com/tag/js/gpt.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://ads.admaru.com/js/huffingtonpost_admaru_incontent_bigbanner.js?t=1548108177791 <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1395}]', u'[console.log([Sticky Ad] > Configuration: {"stickyEnabled":true,"thresholdHeight":109,"stickyDuration":5000,"waitForAdInterval":1500,"waitForAdRetries":150,"offsetFromTop":250}) https://amp.assets.huffpost.com/assets-a380132db9db2f1d3548a6c503316cc8/js/scripts.min.js:2:2905]', u'[JavaScript Warning: "The resource at https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45ab8f20000001016bbb8c.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c0c8230000ef001f9fe0.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45756f360000cb04680c1d.png?cache=7T2hkZGnhY&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45c662240000d9009fa112.jpeg?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c4583f93b00009901689104.jpeg?cache=FoDxZs6gil&ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5c45b6a8200000d4006bbb94.png?ops=150_150 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c0a1a9a2100006708ca2de2.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=d56289fa85f8394c6a49ddf03980316ad33d4261 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5c10a2321d00002c023152f7.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=dcc8bf2c8865a72ae459b6e05b71fd09cb9cd78b was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5bebc3331f0000ff042635ef.png%3Fcache%3DoTI0d0vu3Y%26ops%3D118_67&client=cbc79c14efcebee57402&signature=0448596e1971abf32621c44b58a86f0198952d08 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b6005fd1900002b01500bce.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=b023397ac48c58c06e23f57a56cede516cf6f41f was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b49a2b220000045003730fe.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40cdbca0cf4a933f3db0d589d7a0a81909a4110c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5add9f2c1800002800384493.jpeg%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=40bbfd8344cd5bc109e41c6ea4472a245c70d638 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a84033a1e0000dc007aba7a.jpeg%3Fcache%3D2GNxH5Dg2s%26ops%3D118_67&client=cbc79c14efcebee57402&signature=bf4abef4b04d7ebd55e4d2283a1ad21b1395497c was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5b2dfe722000004200b95363.png%3Fcache%3DuCpgPbxTQf%26ops%3D118_67&client=cbc79c14efcebee57402&signature=b8e41cf32114fb4eb418b9806a94ada61331f944 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a72e08116000020001396f0.jpeg%3Fcache%3DlN5UE5r0Ay%26ops%3D118_67&client=cbc79c14efcebee57402&signature=4b6ba778b24f7c83f441f942a72fb12d973ada7e was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://o.aolcdn.com/images/dims?quality=80&thumbnail=118%2C67&image_uri=https%3A%2F%2Fimg.huffingtonpost.com%2Fasset%2F5a717c7716000020001394a9.png%3Fops%3D118_67&client=cbc79c14efcebee57402&signature=edb1ff7015398d4c6950fef01cfef4d1e530040a was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://connect.facebook.net/en_US/fbevents.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://platform.instagram.com/en_US/embeds.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://img.huffingtonpost.com/asset/5a6a88471d000026006acd90.jpeg?ops=40_40 was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://assets.pinterest.com/js/pinit.js was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "The resource at https://www.youtube.com/iframe_api was blocked because content blocking is enabled." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 0}]', u'[JavaScript Warning: "https://cdn.taboola.com/libtrc/huffingtonpost-korea/loader.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://ad.hani.co.kr/RealMedia/ads/adstream_sx.ads/www.huffingtonpost.kr/a_all@BottomRight <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1815}]', u'[JavaScript Warning: "https://d1z2jf7jlzjs58.cloudfront.net/p.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]', u'[JavaScript Warning: "https://connect.facebook.net/en_US/fbevents.js <script> ." {file: "https://www.huffingtonpost.kr/entry/story_kr_5c3ff38be4b041e98ffa9af2?utm_id=naver" line: 1}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes problem type desktop site instead of mobile site description font does not optimize as mobile site steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel nightly console messages u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u u n u u u u u u u u u u u u u u u u u u u u u u u u u u u u u configuration stickyenabled true thresholdheight stickyduration waitforadinterval waitforadretries offsetfromtop u u u u u u u u u u u u u u u u u u u u u u u u u u u from with ❤️
0
208,613
23,622,841,833
IssuesEvent
2022-08-24 22:45:03
kube-tarian/sigrun
https://api.github.com/repos/kube-tarian/sigrun
closed
CVE-2020-15257 (Medium) detected in github.com/docker/docker-v20.10.7
security vulnerability no-issue-activity
## CVE-2020-15257 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/docker/docker-v20.10.7</b></p></summary> <p>Moby Project - a collaborative project for the container ecosystem to assemble container-based systems</p> <p> Dependency Hierarchy: - github.com/google/go-containerregistry-v0.6.0 (Root Library) - github.com/docker/cli-v20.10.7 - :x: **github.com/docker/docker-v20.10.7** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kube-tarian/sigrun/commit/df1f7d3f67826e841793324e4796be4fbd91c00f">df1f7d3f67826e841793324e4796be4fbd91c00f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> containerd is an industry-standard container runtime and is available as a daemon for Linux and Windows. In containerd before versions 1.3.9 and 1.4.3, the containerd-shim API is improperly exposed to host network containers. Access controls for the shim’s API socket verified that the connecting process had an effective UID of 0, but did not otherwise restrict access to the abstract Unix domain socket. This would allow malicious containers running in the same network namespace as the shim, with an effective UID of 0 but otherwise reduced privileges, to cause new processes to be run with elevated privileges. This vulnerability has been fixed in containerd 1.3.9 and 1.4.3. Users should update to these versions as soon as they are released. It should be noted that containers started with an old version of containerd-shim should be stopped and restarted, as running containers will continue to be vulnerable even after an upgrade. If you are not providing the ability for untrusted users to start containers in the same network namespace as the shim (typically the "host" network namespace, for example with docker run --net=host or hostNetwork: true in a Kubernetes pod) and run with an effective UID of 0, you are not vulnerable to this issue. If you are running containers with a vulnerable configuration, you can deny access to all abstract sockets with AppArmor by adding a line similar to deny unix addr=@**, to your policy. It is best practice to run containers with a reduced set of privileges, with a non-zero UID, and with isolated namespaces. The containerd maintainers strongly advise against sharing namespaces with the host. Reducing the set of isolation mechanisms used for a container necessarily increases that container's privilege, regardless of what container runtime is used for running that container. <p>Publish Date: 2020-12-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15257>CVE-2020-15257</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4">https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4</a></p> <p>Release Date: 2020-12-01</p> <p>Fix Resolution: v1.3.9,v1.4.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15257 (Medium) detected in github.com/docker/docker-v20.10.7 - ## CVE-2020-15257 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/docker/docker-v20.10.7</b></p></summary> <p>Moby Project - a collaborative project for the container ecosystem to assemble container-based systems</p> <p> Dependency Hierarchy: - github.com/google/go-containerregistry-v0.6.0 (Root Library) - github.com/docker/cli-v20.10.7 - :x: **github.com/docker/docker-v20.10.7** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kube-tarian/sigrun/commit/df1f7d3f67826e841793324e4796be4fbd91c00f">df1f7d3f67826e841793324e4796be4fbd91c00f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> containerd is an industry-standard container runtime and is available as a daemon for Linux and Windows. In containerd before versions 1.3.9 and 1.4.3, the containerd-shim API is improperly exposed to host network containers. Access controls for the shim’s API socket verified that the connecting process had an effective UID of 0, but did not otherwise restrict access to the abstract Unix domain socket. This would allow malicious containers running in the same network namespace as the shim, with an effective UID of 0 but otherwise reduced privileges, to cause new processes to be run with elevated privileges. This vulnerability has been fixed in containerd 1.3.9 and 1.4.3. Users should update to these versions as soon as they are released. It should be noted that containers started with an old version of containerd-shim should be stopped and restarted, as running containers will continue to be vulnerable even after an upgrade. If you are not providing the ability for untrusted users to start containers in the same network namespace as the shim (typically the "host" network namespace, for example with docker run --net=host or hostNetwork: true in a Kubernetes pod) and run with an effective UID of 0, you are not vulnerable to this issue. If you are running containers with a vulnerable configuration, you can deny access to all abstract sockets with AppArmor by adding a line similar to deny unix addr=@**, to your policy. It is best practice to run containers with a reduced set of privileges, with a non-zero UID, and with isolated namespaces. The containerd maintainers strongly advise against sharing namespaces with the host. Reducing the set of isolation mechanisms used for a container necessarily increases that container's privilege, regardless of what container runtime is used for running that container. <p>Publish Date: 2020-12-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15257>CVE-2020-15257</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4">https://github.com/containerd/containerd/security/advisories/GHSA-36xw-fx78-c5r4</a></p> <p>Release Date: 2020-12-01</p> <p>Fix Resolution: v1.3.9,v1.4.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in github com docker docker cve medium severity vulnerability vulnerable library github com docker docker moby project a collaborative project for the container ecosystem to assemble container based systems dependency hierarchy github com google go containerregistry root library github com docker cli x github com docker docker vulnerable library found in head commit a href found in base branch main vulnerability details containerd is an industry standard container runtime and is available as a daemon for linux and windows in containerd before versions and the containerd shim api is improperly exposed to host network containers access controls for the shim’s api socket verified that the connecting process had an effective uid of but did not otherwise restrict access to the abstract unix domain socket this would allow malicious containers running in the same network namespace as the shim with an effective uid of but otherwise reduced privileges to cause new processes to be run with elevated privileges this vulnerability has been fixed in containerd and users should update to these versions as soon as they are released it should be noted that containers started with an old version of containerd shim should be stopped and restarted as running containers will continue to be vulnerable even after an upgrade if you are not providing the ability for untrusted users to start containers in the same network namespace as the shim typically the host network namespace for example with docker run net host or hostnetwork true in a kubernetes pod and run with an effective uid of you are not vulnerable to this issue if you are running containers with a vulnerable configuration you can deny access to all abstract sockets with apparmor by adding a line similar to deny unix addr to your policy it is best practice to run containers with a reduced set of privileges with a non zero uid and with isolated namespaces the containerd maintainers strongly advise against sharing namespaces with the host reducing the set of isolation mechanisms used for a container necessarily increases that container s privilege regardless of what container runtime is used for running that container publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
3,255
6,331,988,196
IssuesEvent
2017-07-26 11:27:17
gaocegege/Processing.R
https://api.github.com/repos/gaocegege/Processing.R
closed
Support the event function
community/processing priority/p0 size/no-idea status/WIP type/enhancement
The event function is call when the event has happen. the code is inspired by processing.py ```python value = 0 def draw(): fill(value) rect(25, 25, 50, 50) def mouseClicked(): global value if value == 0: value = 255 else: value = 0 ```
1.0
Support the event function - The event function is call when the event has happen. the code is inspired by processing.py ```python value = 0 def draw(): fill(value) rect(25, 25, 50, 50) def mouseClicked(): global value if value == 0: value = 255 else: value = 0 ```
process
support the event function the event function is call when the event has happen the code is inspired by processing py python value def draw fill value rect def mouseclicked global value if value value else value
1