Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
789
4,389,731,331
IssuesEvent
2016-08-08 23:20:15
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
unarchive doesn't extract changed tar file
bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME unarchive module ansible-modules-core/files/unarchive.py ##### ANSIBLE VERSION 2.1.0.0-1ppa~trusty ##### CONFIGURATION none ##### OS / ENVIRONMENT N/A ##### SUMMARY unarchive is not updating dest (extract the tar) if tar content changed and still keep old file structures, (means if you run "tar -d", it reports Mod Time changed) by checking the ansible-modules-core/files/unarchive.py, it seems TgzArchive is not handling "Mod time differs" condition in is_unarchived method ##### STEPS TO REPRODUCE 1) prepare tar file from folder, make it demo.tar 2) run this command in playbook - unarchive: src=/opt/packages/demo.tar dest=/opt/app/ copy=no 3) update content of any file, make newer demo.tar 4) run the same playbook ##### EXPECTED RESULTS the dest folder should be updated, ##### ACTUAL RESULTS ansible ubuntu skipped it, even though the tar file is changed TASK [unarchive] *************************************************************** ok: [localhost]
True
unarchive doesn't extract changed tar file - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME unarchive module ansible-modules-core/files/unarchive.py ##### ANSIBLE VERSION 2.1.0.0-1ppa~trusty ##### CONFIGURATION none ##### OS / ENVIRONMENT N/A ##### SUMMARY unarchive is not updating dest (extract the tar) if tar content changed and still keep old file structures, (means if you run "tar -d", it reports Mod Time changed) by checking the ansible-modules-core/files/unarchive.py, it seems TgzArchive is not handling "Mod time differs" condition in is_unarchived method ##### STEPS TO REPRODUCE 1) prepare tar file from folder, make it demo.tar 2) run this command in playbook - unarchive: src=/opt/packages/demo.tar dest=/opt/app/ copy=no 3) update content of any file, make newer demo.tar 4) run the same playbook ##### EXPECTED RESULTS the dest folder should be updated, ##### ACTUAL RESULTS ansible ubuntu skipped it, even though the tar file is changed TASK [unarchive] *************************************************************** ok: [localhost]
main
unarchive doesn t extract changed tar file issue type bug report component name unarchive module ansible modules core files unarchive py ansible version trusty configuration none os environment n a summary unarchive is not updating dest extract the tar if tar content changed and still keep old file structures means if you run tar d it reports mod time changed by checking the ansible modules core files unarchive py it seems tgzarchive is not handling mod time differs condition in is unarchived method steps to reproduce prepare tar file from folder make it demo tar run this command in playbook unarchive src opt packages demo tar dest opt app copy no update content of any file make newer demo tar run the same playbook expected results the dest folder should be updated actual results ansible ubuntu skipped it even though the tar file is changed task ok
1
185,990
21,902,520,607
IssuesEvent
2022-05-20 14:41:09
CMSgov/mdct-mcr
https://api.github.com/repos/CMSgov/mdct-mcr
opened
SHF - mdct-mcr - main - MEDIUM - CloudFront.7 CloudFront distributions should use custom SSL/TLS certificates
security-hub main
************************************************************** __This issue was generated from Security Hub data and is managed through automation.__ Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome. Finding Id: arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/CloudFront.7/finding/fc83ba38-b50c-4cab-bc97-6fb9bcc80843 ************************************************************** ## Type of Issue: - [x] Security Hub Finding ## Title: CloudFront.7 CloudFront distributions should use custom SSL/TLS certificates ## Id: arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/CloudFront.7/finding/fc83ba38-b50c-4cab-bc97-6fb9bcc80843 (You may use this ID to lookup this finding's details in Security Hub) ## Description This control checks if CloudFront distributions are using the default SSL/TLS certificate CloudFront provides instead of a custom one. This control fails for a CloudFront distribution if it uses the default SSL/TLS certificate. ## Remediation https://docs.aws.amazon.com/console/securityhub/CloudFront.7/remediation ## AC: - The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
True
SHF - mdct-mcr - main - MEDIUM - CloudFront.7 CloudFront distributions should use custom SSL/TLS certificates - ************************************************************** __This issue was generated from Security Hub data and is managed through automation.__ Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome. Finding Id: arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/CloudFront.7/finding/fc83ba38-b50c-4cab-bc97-6fb9bcc80843 ************************************************************** ## Type of Issue: - [x] Security Hub Finding ## Title: CloudFront.7 CloudFront distributions should use custom SSL/TLS certificates ## Id: arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/CloudFront.7/finding/fc83ba38-b50c-4cab-bc97-6fb9bcc80843 (You may use this ID to lookup this finding's details in Security Hub) ## Description This control checks if CloudFront distributions are using the default SSL/TLS certificate CloudFront provides instead of a custom one. This control fails for a CloudFront distribution if it uses the default SSL/TLS certificate. ## Remediation https://docs.aws.amazon.com/console/securityhub/CloudFront.7/remediation ## AC: - The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
non_main
shf mdct mcr main medium cloudfront cloudfront distributions should use custom ssl tls certificates this issue was generated from security hub data and is managed through automation please do not edit the title or body of this issue or remove the security hub tag all other edits comments are welcome finding id arn aws securityhub us east subscription aws foundational security best practices v cloudfront finding type of issue security hub finding title cloudfront cloudfront distributions should use custom ssl tls certificates id arn aws securityhub us east subscription aws foundational security best practices v cloudfront finding you may use this id to lookup this finding s details in security hub description this control checks if cloudfront distributions are using the default ssl tls certificate cloudfront provides instead of a custom one this control fails for a cloudfront distribution if it uses the default ssl tls certificate remediation ac the security hub finding is resolved or suppressed indicated by a workflow status of resolved or suppressed
0
39,598
6,758,120,588
IssuesEvent
2017-10-24 13:17:14
metoppv/improver
https://api.github.com/repos/metoppv/improver
closed
Decide on and implement a metadata convention for weather symbols
Type: Documentation Type: Feature
As an IMPROVER developer, I want to be able to store weather symbols within a netCDF file in a consistent way so that they can be read in and interpreted correctly. Related issues: #292 This issue should : - Document the existing metadata convention using the existing weather symbol relationships, for example, 0 = Clear night, 1 = Sunny day, 2 = Partly cloudy night. - Create a utility that applies this convention. This utility will be used for creating a cube containing weather symbol information, such that the values that have been attributed to a particular site or grid point, can be cross-referenced to the meaning of the weather symbol easily. Acceptance criteria: * Document the existing metadata convention for weather symbol codes and their description in Confluence. * Create a utility that can be used for creating a cube and applying the metadata convention. * **New criteria: Generate sample data for verification, over several cycles.**
1.0
Decide on and implement a metadata convention for weather symbols - As an IMPROVER developer, I want to be able to store weather symbols within a netCDF file in a consistent way so that they can be read in and interpreted correctly. Related issues: #292 This issue should : - Document the existing metadata convention using the existing weather symbol relationships, for example, 0 = Clear night, 1 = Sunny day, 2 = Partly cloudy night. - Create a utility that applies this convention. This utility will be used for creating a cube containing weather symbol information, such that the values that have been attributed to a particular site or grid point, can be cross-referenced to the meaning of the weather symbol easily. Acceptance criteria: * Document the existing metadata convention for weather symbol codes and their description in Confluence. * Create a utility that can be used for creating a cube and applying the metadata convention. * **New criteria: Generate sample data for verification, over several cycles.**
non_main
decide on and implement a metadata convention for weather symbols as an improver developer i want to be able to store weather symbols within a netcdf file in a consistent way so that they can be read in and interpreted correctly related issues this issue should document the existing metadata convention using the existing weather symbol relationships for example clear night sunny day partly cloudy night create a utility that applies this convention this utility will be used for creating a cube containing weather symbol information such that the values that have been attributed to a particular site or grid point can be cross referenced to the meaning of the weather symbol easily acceptance criteria document the existing metadata convention for weather symbol codes and their description in confluence create a utility that can be used for creating a cube and applying the metadata convention new criteria generate sample data for verification over several cycles
0
1,893
6,577,538,313
IssuesEvent
2017-09-12 01:36:51
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Please provide yum swap capability
affects_2.0 feature_idea waiting_on_maintainer
##### Issue Type: Feature Idea ##### Plugin Name: yum ##### Ansible Version: 2.0.1.0 ##### Environment: N/A ##### Summary: yum module does not provide swap capability
True
Please provide yum swap capability - ##### Issue Type: Feature Idea ##### Plugin Name: yum ##### Ansible Version: 2.0.1.0 ##### Environment: N/A ##### Summary: yum module does not provide swap capability
main
please provide yum swap capability issue type feature idea plugin name yum ansible version environment n a summary yum module does not provide swap capability
1
1,161
5,053,282,740
IssuesEvent
2016-12-21 07:25:23
betaflight/betaflight
https://api.github.com/repos/betaflight/betaflight
closed
ARF3 DSHOT can not work
For Target Maintainer
arf3 3.1 dshot can not work According to f3evo re-made the target but Motor2 not worl in dshot,multishot is ok target.c: `#include <stdint.h> #include <platform.h> #include "drivers/io.h" #include "drivers/timer.h" #include "drivers/timer_def.h" #include "drivers/dma.h" const timerHardware_t timerHardware[USABLE_TIMER_CHANNEL_COUNT] = { // PPM / UART2 RX DEF_TIM(TIM8, CH1, PA15, TIM_USE_PPM, 0 ), // PPM DEF_TIM(TIM3, CH4, PB1, TIM_USE_MOTOR, 1 ), // PWM1 DEF_TIM(TIM3, CH2, PA7, TIM_USE_MOTOR, 1 ), // PWM2 DEF_TIM(TIM2, CH3, PA2, TIM_USE_MOTOR, 1 ), // PWM3 DEF_TIM(TIM2, CH2, PA1, TIM_USE_MOTOR, 1 ), // PWM4 DEF_TIM(TIM3, CH3, PB0, TIM_USE_MOTOR, 1 ), // PWM5 DEF_TIM(TIM3, CH1, PA6, TIM_USE_MOTOR, 1 ), // PWM6 DEF_TIM(TIM15, CH2, PA3, TIM_USE_MOTOR, 1 ), // PWM7 DEF_TIM(TIM2, CH1, PA0, TIM_USE_MOTOR, 1 ), // PWM8 DEF_TIM(TIM2, CH3, PB10, TIM_USE_MOTOR, 1 ), DEF_TIM(TIM2, CH4, PB11, TIM_USE_MOTOR, 1 ), DEF_TIM(TIM1, CH1, PA8, TIM_USE_LED | TIM_USE_TRANSPONDER, 1 ), // LED_STRIP / TRANSPONDER };` target.h: ` #pragma once #define TARGET_BOARD_IDENTIFIER "ARF3" #define CONFIG_FASTLOOP_PREFERRED_ACC ACC_DEFAULT #define LED0 PB8 #define BEEPER PC15 #define BEEPER_INVERTED #define USE_EXTI #define MPU_INT_EXTI PC13 #define EXTI15_10_CALLBACK_HANDLER_COUNT 2 // MPU_INT, SDCardDetect #define USE_MPU_DATA_READY_SIGNAL #define ENSURE_MPU_DATA_READY_IS_LOW #define USE_MAG_DATA_READY_SIGNAL #define ENSURE_MAG_DATA_READY_IS_HIGH #define USE_ESC_TELEMETRY #define GYRO #define USE_GYRO_SPI_MPU6500 #define ACC #define USE_ACC_SPI_MPU6500 #define ACC_MPU6500_ALIGN CW180_DEG #define GYRO_MPU6500_ALIGN CW180_DEG #define BARO #define USE_BARO_BMP280 #define MAG #define USE_MAG_AK8963 //#define USE_MAG_HMC5883 // External #define MAG_AK8963_ALIGN CW90_DEG_FLIP //#define SONAR #define USB_IO #define USE_VCP #define USE_UART1 #define USE_UART2 #define USE_UART3 #define SERIAL_PORT_COUNT 4 #define USE_ESCSERIAL #define ESCSERIAL_TIMER_TX_HARDWARE 0 // PWM 1 #define UART1_TX_PIN PA9 #define UART1_RX_PIN PA10 #define UART2_TX_PIN PA14 // PA14 / SWCLK #define UART2_RX_PIN PA15 #define UART3_TX_PIN PB10 // PB10 (AF7) #define UART3_RX_PIN PB11 // PB11 (AF7) #define USE_I2C #define I2C_DEVICE (I2CDEV_1) // PB6/SCL, PB7/SDA #define USE_SPI #define USE_SPI_DEVICE_1 // PB9,3,4,5 on AF5 SPI1 (MPU) #define USE_SPI_DEVICE_2 // PB12,13,14,15 on AF5 SPI2 (SDCard) #define SPI1_NSS_PIN PB9 #define SPI1_SCK_PIN PB3 #define SPI1_MISO_PIN PB4 #define SPI1_MOSI_PIN PB5 #define SPI2_NSS_PIN PB12 #define SPI2_SCK_PIN PB13 #define SPI2_MISO_PIN PB14 #define SPI2_MOSI_PIN PB15 #define USE_SDCARD #define USE_SDCARD_SPI2 #define SDCARD_DETECT_INVERTED #define SDCARD_DETECT_PIN PC14 #define SDCARD_SPI_INSTANCE SPI2 #define SDCARD_SPI_CS_PIN SPI2_NSS_PIN // SPI2 is on the APB1 bus whose clock runs at 36MHz. Divide to under 400kHz for init: #define SDCARD_SPI_INITIALIZATION_CLOCK_DIVIDER 128 // Divide to under 25MHz for normal operation: #define SDCARD_SPI_FULL_SPEED_CLOCK_DIVIDER 2 // Note, this is the same DMA channel as UART1_RX. Luckily we don't use DMA for USART Rx. #define SDCARD_DMA_CHANNEL_TX DMA1_Channel5 #define SDCARD_DMA_CHANNEL_TX_COMPLETE_FLAG DMA1_FLAG_TC5 #define MPU6500_CS_PIN PB9 #define MPU6500_SPI_INSTANCE SPI1 #define BOARD_HAS_VOLTAGE_DIVIDER #define USE_ADC #define ADC_INSTANCE ADC2 #define VBAT_ADC_PIN PA5 #define CURRENT_METER_ADC_PIN PA4 #define RSSI_ADC_PIN PB2 #define LED_STRIP #define TRANSPONDER #define ENABLE_BLACKBOX_LOGGING_ON_SDCARD_BY_DEFAULT #define DEFAULT_RX_FEATURE FEATURE_RX_PPM #define DEFAULT_FEATURES (FEATURE_TRANSPONDER | FEATURE_BLACKBOX | FEATURE_RSSI_ADC | FEATURE_CURRENT_METER | FEATURE_TELEMETRY) #define SPEKTRUM_BIND // USART3, #define BIND_PIN PB11 #define USE_SERIAL_4WAY_BLHELI_INTERFACE // IO - stm32f303cc in 48pin package #define TARGET_IO_PORTA 0xffff #define TARGET_IO_PORTB 0xffff #define TARGET_IO_PORTC (BIT(13)|BIT(14)|BIT(15)) #define TARGET_IO_PORTF (BIT(0)|BIT(1)|BIT(4)) #define USABLE_TIMER_CHANNEL_COUNT 12 // PPM, 8 PWM, UART3 RX/TX, LED Strip #define USED_TIMERS (TIM_N(1) | TIM_N(2) | TIM_N(3) | TIM_N(8) | TIM_N(15)) `
True
ARF3 DSHOT can not work - arf3 3.1 dshot can not work According to f3evo re-made the target but Motor2 not worl in dshot,multishot is ok target.c: `#include <stdint.h> #include <platform.h> #include "drivers/io.h" #include "drivers/timer.h" #include "drivers/timer_def.h" #include "drivers/dma.h" const timerHardware_t timerHardware[USABLE_TIMER_CHANNEL_COUNT] = { // PPM / UART2 RX DEF_TIM(TIM8, CH1, PA15, TIM_USE_PPM, 0 ), // PPM DEF_TIM(TIM3, CH4, PB1, TIM_USE_MOTOR, 1 ), // PWM1 DEF_TIM(TIM3, CH2, PA7, TIM_USE_MOTOR, 1 ), // PWM2 DEF_TIM(TIM2, CH3, PA2, TIM_USE_MOTOR, 1 ), // PWM3 DEF_TIM(TIM2, CH2, PA1, TIM_USE_MOTOR, 1 ), // PWM4 DEF_TIM(TIM3, CH3, PB0, TIM_USE_MOTOR, 1 ), // PWM5 DEF_TIM(TIM3, CH1, PA6, TIM_USE_MOTOR, 1 ), // PWM6 DEF_TIM(TIM15, CH2, PA3, TIM_USE_MOTOR, 1 ), // PWM7 DEF_TIM(TIM2, CH1, PA0, TIM_USE_MOTOR, 1 ), // PWM8 DEF_TIM(TIM2, CH3, PB10, TIM_USE_MOTOR, 1 ), DEF_TIM(TIM2, CH4, PB11, TIM_USE_MOTOR, 1 ), DEF_TIM(TIM1, CH1, PA8, TIM_USE_LED | TIM_USE_TRANSPONDER, 1 ), // LED_STRIP / TRANSPONDER };` target.h: ` #pragma once #define TARGET_BOARD_IDENTIFIER "ARF3" #define CONFIG_FASTLOOP_PREFERRED_ACC ACC_DEFAULT #define LED0 PB8 #define BEEPER PC15 #define BEEPER_INVERTED #define USE_EXTI #define MPU_INT_EXTI PC13 #define EXTI15_10_CALLBACK_HANDLER_COUNT 2 // MPU_INT, SDCardDetect #define USE_MPU_DATA_READY_SIGNAL #define ENSURE_MPU_DATA_READY_IS_LOW #define USE_MAG_DATA_READY_SIGNAL #define ENSURE_MAG_DATA_READY_IS_HIGH #define USE_ESC_TELEMETRY #define GYRO #define USE_GYRO_SPI_MPU6500 #define ACC #define USE_ACC_SPI_MPU6500 #define ACC_MPU6500_ALIGN CW180_DEG #define GYRO_MPU6500_ALIGN CW180_DEG #define BARO #define USE_BARO_BMP280 #define MAG #define USE_MAG_AK8963 //#define USE_MAG_HMC5883 // External #define MAG_AK8963_ALIGN CW90_DEG_FLIP //#define SONAR #define USB_IO #define USE_VCP #define USE_UART1 #define USE_UART2 #define USE_UART3 #define SERIAL_PORT_COUNT 4 #define USE_ESCSERIAL #define ESCSERIAL_TIMER_TX_HARDWARE 0 // PWM 1 #define UART1_TX_PIN PA9 #define UART1_RX_PIN PA10 #define UART2_TX_PIN PA14 // PA14 / SWCLK #define UART2_RX_PIN PA15 #define UART3_TX_PIN PB10 // PB10 (AF7) #define UART3_RX_PIN PB11 // PB11 (AF7) #define USE_I2C #define I2C_DEVICE (I2CDEV_1) // PB6/SCL, PB7/SDA #define USE_SPI #define USE_SPI_DEVICE_1 // PB9,3,4,5 on AF5 SPI1 (MPU) #define USE_SPI_DEVICE_2 // PB12,13,14,15 on AF5 SPI2 (SDCard) #define SPI1_NSS_PIN PB9 #define SPI1_SCK_PIN PB3 #define SPI1_MISO_PIN PB4 #define SPI1_MOSI_PIN PB5 #define SPI2_NSS_PIN PB12 #define SPI2_SCK_PIN PB13 #define SPI2_MISO_PIN PB14 #define SPI2_MOSI_PIN PB15 #define USE_SDCARD #define USE_SDCARD_SPI2 #define SDCARD_DETECT_INVERTED #define SDCARD_DETECT_PIN PC14 #define SDCARD_SPI_INSTANCE SPI2 #define SDCARD_SPI_CS_PIN SPI2_NSS_PIN // SPI2 is on the APB1 bus whose clock runs at 36MHz. Divide to under 400kHz for init: #define SDCARD_SPI_INITIALIZATION_CLOCK_DIVIDER 128 // Divide to under 25MHz for normal operation: #define SDCARD_SPI_FULL_SPEED_CLOCK_DIVIDER 2 // Note, this is the same DMA channel as UART1_RX. Luckily we don't use DMA for USART Rx. #define SDCARD_DMA_CHANNEL_TX DMA1_Channel5 #define SDCARD_DMA_CHANNEL_TX_COMPLETE_FLAG DMA1_FLAG_TC5 #define MPU6500_CS_PIN PB9 #define MPU6500_SPI_INSTANCE SPI1 #define BOARD_HAS_VOLTAGE_DIVIDER #define USE_ADC #define ADC_INSTANCE ADC2 #define VBAT_ADC_PIN PA5 #define CURRENT_METER_ADC_PIN PA4 #define RSSI_ADC_PIN PB2 #define LED_STRIP #define TRANSPONDER #define ENABLE_BLACKBOX_LOGGING_ON_SDCARD_BY_DEFAULT #define DEFAULT_RX_FEATURE FEATURE_RX_PPM #define DEFAULT_FEATURES (FEATURE_TRANSPONDER | FEATURE_BLACKBOX | FEATURE_RSSI_ADC | FEATURE_CURRENT_METER | FEATURE_TELEMETRY) #define SPEKTRUM_BIND // USART3, #define BIND_PIN PB11 #define USE_SERIAL_4WAY_BLHELI_INTERFACE // IO - stm32f303cc in 48pin package #define TARGET_IO_PORTA 0xffff #define TARGET_IO_PORTB 0xffff #define TARGET_IO_PORTC (BIT(13)|BIT(14)|BIT(15)) #define TARGET_IO_PORTF (BIT(0)|BIT(1)|BIT(4)) #define USABLE_TIMER_CHANNEL_COUNT 12 // PPM, 8 PWM, UART3 RX/TX, LED Strip #define USED_TIMERS (TIM_N(1) | TIM_N(2) | TIM_N(3) | TIM_N(8) | TIM_N(15)) `
main
dshot can not work dshot can not work according to re made the target but not worl in dshot multishot is ok target c include include include drivers io h include drivers timer h include drivers timer def h include drivers dma h const timerhardware t timerhardware ppm rx def tim tim use ppm ppm def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use motor def tim tim use led tim use transponder led strip transponder target h pragma once define target board identifier define config fastloop preferred acc acc default define define beeper define beeper inverted define use exti define mpu int exti define callback handler count mpu int sdcarddetect define use mpu data ready signal define ensure mpu data ready is low define use mag data ready signal define ensure mag data ready is high define use esc telemetry define gyro define use gyro spi define acc define use acc spi define acc align deg define gyro align deg define baro define use baro define mag define use mag define use mag external define mag align deg flip define sonar define usb io define use vcp define use define use define use define serial port count define use escserial define escserial timer tx hardware pwm define tx pin define rx pin define tx pin swclk define rx pin define tx pin define rx pin define use define device scl sda define use spi define use spi device on mpu define use spi device on sdcard define nss pin define sck pin define miso pin define mosi pin define nss pin define sck pin define miso pin define mosi pin define use sdcard define use sdcard define sdcard detect inverted define sdcard detect pin define sdcard spi instance define sdcard spi cs pin nss pin is on the bus whose clock runs at divide to under for init define sdcard spi initialization clock divider divide to under for normal operation define sdcard spi full speed clock divider note this is the same dma channel as rx luckily we don t use dma for usart rx define sdcard dma channel tx define sdcard dma channel tx complete flag flag define cs pin define spi instance define board has voltage divider define use adc define adc instance define vbat adc pin define current meter adc pin define rssi adc pin define led strip define transponder define enable blackbox logging on sdcard by default define default rx feature feature rx ppm define default features feature transponder feature blackbox feature rssi adc feature current meter feature telemetry define spektrum bind define bind pin define use serial blheli interface io in package define target io porta define target io portb define target io portc bit bit bit define target io portf bit bit bit define usable timer channel count ppm pwm rx tx led strip define used timers tim n tim n tim n tim n tim n
1
1,852
6,577,396,441
IssuesEvent
2017-09-12 00:37:22
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
The ios_config module does not delete the username from the router
affects_2.0 bug_report networking waiting_on_maintainer
##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> networking/ios_config ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.0.1.0 ``` ##### CONFIGURATION No configuration. ##### OS / ENVIRONMENT Working on Ubuntu, release trusty. The issue is not platform dependent. ##### SUMMARY When using command "no username" with ios_config module, the task timeouts. ##### STEPS TO REPRODUCE I am using the following task configuration in the playbook. ``` - name: Delete users ios_config: host: "{{ ansible_ssh_host }}" username: "{{ username }}" password: "{{ password }}" lines: - no username admin ``` The result of the task is: ``` TASK [Delete users] ************************************************************ fatal: [hub]: FAILED! => {"changed": false, "commands": ["configure terminal", "no username admin"], "failed": true, "msg": "timeout trying to send command"} ``` The reason for this is that Cisco ask for confirmation when deleting the username from the configuration. ``` R2(config)#no username admin This operation will remove all username related configurations with same name.Do you want to continue? [confirm] R2(config)# ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS It is expected that command should be deleted from the configuration. ##### ACTUAL RESULTS ``` fatal: [wan1]: FAILED! => {"changed": false, "commands": ["configure terminal", "no username admin"], "failed": true, "invocation": {"module_args": {"after": null, "auth_pass": null, "authorize": false, "before": null, "config": null, "force": false, "host": "192.168.35.152", "lines": ["no username admin"], "match": "line", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": 22, "provider": null, "replace": "line", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"}, "module_name": "ios_config"}, "msg": "timeout trying to send command"} ```
True
The ios_config module does not delete the username from the router - ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> networking/ios_config ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.0.1.0 ``` ##### CONFIGURATION No configuration. ##### OS / ENVIRONMENT Working on Ubuntu, release trusty. The issue is not platform dependent. ##### SUMMARY When using command "no username" with ios_config module, the task timeouts. ##### STEPS TO REPRODUCE I am using the following task configuration in the playbook. ``` - name: Delete users ios_config: host: "{{ ansible_ssh_host }}" username: "{{ username }}" password: "{{ password }}" lines: - no username admin ``` The result of the task is: ``` TASK [Delete users] ************************************************************ fatal: [hub]: FAILED! => {"changed": false, "commands": ["configure terminal", "no username admin"], "failed": true, "msg": "timeout trying to send command"} ``` The reason for this is that Cisco ask for confirmation when deleting the username from the configuration. ``` R2(config)#no username admin This operation will remove all username related configurations with same name.Do you want to continue? [confirm] R2(config)# ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS It is expected that command should be deleted from the configuration. ##### ACTUAL RESULTS ``` fatal: [wan1]: FAILED! => {"changed": false, "commands": ["configure terminal", "no username admin"], "failed": true, "invocation": {"module_args": {"after": null, "auth_pass": null, "authorize": false, "before": null, "config": null, "force": false, "host": "192.168.35.152", "lines": ["no username admin"], "match": "line", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": 22, "provider": null, "replace": "line", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"}, "module_name": "ios_config"}, "msg": "timeout trying to send command"} ```
main
the ios config module does not delete the username from the router issue type bug report component name networking ios config ansible version ansible version ansible configuration no configuration os environment working on ubuntu release trusty the issue is not platform dependent summary when using command no username with ios config module the task timeouts steps to reproduce i am using the following task configuration in the playbook name delete users ios config host ansible ssh host username username password password lines no username admin the result of the task is task fatal failed changed false commands failed true msg timeout trying to send command the reason for this is that cisco ask for confirmation when deleting the username from the configuration config no username admin this operation will remove all username related configurations with same name do you want to continue config expected results it is expected that command should be deleted from the configuration actual results fatal failed changed false commands failed true invocation module args after null auth pass null authorize false before null config null force false host lines match line parents null password value specified in no log parameter port provider null replace line username value specified in no log parameter module name ios config msg timeout trying to send command
1
502,614
14,562,924,625
IssuesEvent
2020-12-17 01:13:35
kubernetes/website
https://api.github.com/repos/kubernetes/website
closed
Kubernetes v1.19 - Doc Reference Page has mismatch links
kind/bug priority/important-soon triage/accepted
Problem : In the v1.19 version of the Kubernetes site - https://v1-19.docs.kubernetes.io/docs/reference/, we are seeing the link for Kubernetes API - https://v1-19.docs.kubernetes.io/docs/reference/generated/kubernetes-api/v1.20/ Steps to Reproduce : 1. Navigate to the URL: https://kubernetes.io/ and in the versions choose v1.19 2.It will be redirected to https://v1-19.docs.kubernetes.io/ 3.Documentation -> reference, under API Reference section we are able to spot a wrong link for v1.20 _Kubernetes API Reference v1.20_ Solution : https://v1-19.docs.kubernetes.io/docs/reference/generated/kubernetes-api/v1.19/ is the correct Kubernetes API Reference for v1.19 and the link has to be re-titled as **Kubernetes API Reference v1.19** Note: As the further older doc versions are not maintained, not sure if it has to be fixed in those as well.
1.0
Kubernetes v1.19 - Doc Reference Page has mismatch links - Problem : In the v1.19 version of the Kubernetes site - https://v1-19.docs.kubernetes.io/docs/reference/, we are seeing the link for Kubernetes API - https://v1-19.docs.kubernetes.io/docs/reference/generated/kubernetes-api/v1.20/ Steps to Reproduce : 1. Navigate to the URL: https://kubernetes.io/ and in the versions choose v1.19 2.It will be redirected to https://v1-19.docs.kubernetes.io/ 3.Documentation -> reference, under API Reference section we are able to spot a wrong link for v1.20 _Kubernetes API Reference v1.20_ Solution : https://v1-19.docs.kubernetes.io/docs/reference/generated/kubernetes-api/v1.19/ is the correct Kubernetes API Reference for v1.19 and the link has to be re-titled as **Kubernetes API Reference v1.19** Note: As the further older doc versions are not maintained, not sure if it has to be fixed in those as well.
non_main
kubernetes doc reference page has mismatch links problem in the version of the kubernetes site we are seeing the link for kubernetes api steps to reproduce navigate to the url and in the versions choose it will be redirected to documentation reference under api reference section we are able to spot a wrong link for kubernetes api reference solution is the correct kubernetes api reference for and the link has to be re titled as kubernetes api reference note as the further older doc versions are not maintained not sure if it has to be fixed in those as well
0
1,668
6,574,071,002
IssuesEvent
2017-09-11 11:21:11
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
unarchive: unhelpful error message when "owner" does not exist and unarchiving tar.gz
affects_2.3 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME unarchive ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> N/A ##### SUMMARY <!--- Explain the problem briefly --> When "owner" parameter is set to user that does not exist, and we are extracting tar.gz archive the module fails with unrelated error message: "Unexpected error when accessing exploded file: [Errno 2] No such file or directory:" Interesting note: when the file being archived has no extension, ie "test" instead of "test.txt" below, the issue does not occur. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` mkdir /tmp/unarchive_bug cd /tmp/unarchive_bug mkdir unarchived echo "test" > test.txt tar cf testfile.tar test.txt gzip -9 testfile.tar cat <<'EOF' > unarchive_test.yml - hosts: localhost tasks: - name: unarchive test unarchive: copy: no src: /tmp/unarchive_bug/testfile.tar.gz dest: /tmp/unarchive_bug/unarchived owner: doesnotexist EOF ansible-playbook -c local unarchive_test.yml ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Failure message that indicates that user "doesnotexist" does not exist. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` TASK [unarchive test] ********************************************************************************************************************** fatal: [localhost]: FAILED! => {"changed": false, "dest": "/tmp/unarchive_bug/unarchived", "failed": true, "gid": 0, "group": "root", "handler": "TgzArchive", "mode": "0755", "msg": "Unexpected error when accessing exploded file: [Errno 2] No such file or directory: '/tmp/unarchive_bug/unarchived/test.txt'", "owner": "root", "size": 21, "src": "/tmp/unarchive_bug/testfile.tar.gz", "state": "directory", "uid": 0} to retry, use: --limit @/tmp/unarchive_bug/unarchive_test.retry ```
True
unarchive: unhelpful error message when "owner" does not exist and unarchiving tar.gz - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME unarchive ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> N/A ##### SUMMARY <!--- Explain the problem briefly --> When "owner" parameter is set to user that does not exist, and we are extracting tar.gz archive the module fails with unrelated error message: "Unexpected error when accessing exploded file: [Errno 2] No such file or directory:" Interesting note: when the file being archived has no extension, ie "test" instead of "test.txt" below, the issue does not occur. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` mkdir /tmp/unarchive_bug cd /tmp/unarchive_bug mkdir unarchived echo "test" > test.txt tar cf testfile.tar test.txt gzip -9 testfile.tar cat <<'EOF' > unarchive_test.yml - hosts: localhost tasks: - name: unarchive test unarchive: copy: no src: /tmp/unarchive_bug/testfile.tar.gz dest: /tmp/unarchive_bug/unarchived owner: doesnotexist EOF ansible-playbook -c local unarchive_test.yml ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Failure message that indicates that user "doesnotexist" does not exist. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` TASK [unarchive test] ********************************************************************************************************************** fatal: [localhost]: FAILED! => {"changed": false, "dest": "/tmp/unarchive_bug/unarchived", "failed": true, "gid": 0, "group": "root", "handler": "TgzArchive", "mode": "0755", "msg": "Unexpected error when accessing exploded file: [Errno 2] No such file or directory: '/tmp/unarchive_bug/unarchived/test.txt'", "owner": "root", "size": 21, "src": "/tmp/unarchive_bug/testfile.tar.gz", "state": "directory", "uid": 0} to retry, use: --limit @/tmp/unarchive_bug/unarchive_test.retry ```
main
unarchive unhelpful error message when owner does not exist and unarchiving tar gz issue type bug report component name unarchive ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific n a summary when owner parameter is set to user that does not exist and we are extracting tar gz archive the module fails with unrelated error message unexpected error when accessing exploded file no such file or directory interesting note when the file being archived has no extension ie test instead of test txt below the issue does not occur steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used mkdir tmp unarchive bug cd tmp unarchive bug mkdir unarchived echo test test txt tar cf testfile tar test txt gzip testfile tar cat unarchive test yml hosts localhost tasks name unarchive test unarchive copy no src tmp unarchive bug testfile tar gz dest tmp unarchive bug unarchived owner doesnotexist eof ansible playbook c local unarchive test yml expected results failure message that indicates that user doesnotexist does not exist actual results task fatal failed changed false dest tmp unarchive bug unarchived failed true gid group root handler tgzarchive mode msg unexpected error when accessing exploded file no such file or directory tmp unarchive bug unarchived test txt owner root size src tmp unarchive bug testfile tar gz state directory uid to retry use limit tmp unarchive bug unarchive test retry
1
74,016
14,168,251,284
IssuesEvent
2020-11-12 11:27:08
AlmasB/FXGL
https://api.github.com/repos/AlmasB/FXGL
opened
Update documentation for all modules
help wanted medium type:non code
Review all modules and their public API and update/add where documentation is unclear/missing. To make this task more practically feasible, it would be good to provide all modules here and "check-box" them once they have been reviewed. This way, each contributor only needs to review one module, which is a significantly less complex task.
1.0
Update documentation for all modules - Review all modules and their public API and update/add where documentation is unclear/missing. To make this task more practically feasible, it would be good to provide all modules here and "check-box" them once they have been reviewed. This way, each contributor only needs to review one module, which is a significantly less complex task.
non_main
update documentation for all modules review all modules and their public api and update add where documentation is unclear missing to make this task more practically feasible it would be good to provide all modules here and check box them once they have been reviewed this way each contributor only needs to review one module which is a significantly less complex task
0
659,208
21,919,499,274
IssuesEvent
2022-05-22 10:58:20
googleapis/java-bigtable
https://api.github.com/repos/googleapis/java-bigtable
opened
bigtable.data.v2.stub.DynamicFlowControlCallableTest: testIncreasingThreshold failed
priority: p2 type: bug flakybot: issue
Note: #1242 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 8a2abb5b62b7795a5750c194cea5c94fb96878f5 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f79488b8-d824-4852-8219-bec6e348375a), [Sponge](http://sponge2/f79488b8-d824-4852-8219-bec6e348375a) status: failed <details><summary>Test output</summary><br><pre>expected not to be: null at com.google.cloud.bigtable.data.v2.stub.DynamicFlowControlCallableTest.createFlowControlEvent(DynamicFlowControlCallableTest.java:300) at com.google.cloud.bigtable.data.v2.stub.DynamicFlowControlCallableTest.testIncreasingThreshold(DynamicFlowControlCallableTest.java:179) </pre></details>
1.0
bigtable.data.v2.stub.DynamicFlowControlCallableTest: testIncreasingThreshold failed - Note: #1242 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 8a2abb5b62b7795a5750c194cea5c94fb96878f5 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f79488b8-d824-4852-8219-bec6e348375a), [Sponge](http://sponge2/f79488b8-d824-4852-8219-bec6e348375a) status: failed <details><summary>Test output</summary><br><pre>expected not to be: null at com.google.cloud.bigtable.data.v2.stub.DynamicFlowControlCallableTest.createFlowControlEvent(DynamicFlowControlCallableTest.java:300) at com.google.cloud.bigtable.data.v2.stub.DynamicFlowControlCallableTest.testIncreasingThreshold(DynamicFlowControlCallableTest.java:179) </pre></details>
non_main
bigtable data stub dynamicflowcontrolcallabletest testincreasingthreshold failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output expected not to be null at com google cloud bigtable data stub dynamicflowcontrolcallabletest createflowcontrolevent dynamicflowcontrolcallabletest java at com google cloud bigtable data stub dynamicflowcontrolcallabletest testincreasingthreshold dynamicflowcontrolcallabletest java
0
4,295
21,657,891,053
IssuesEvent
2022-05-06 15:50:23
ipfs/ipfs-docs
https://api.github.com/repos/ipfs/ipfs-docs
closed
Document PubSub changes from go-ipfs v0.11
dif/easy good first issue P1 kind/bug need/maintainers-input effort/days
This describes the issue very well: https://github.com/ipfs/go-ipfs/issues/8612 Caused by: https://github.com/ipfs/go-ipfs/issues/8343 Relevant areas: https://docs.ipfs.io/reference/http/api/#api-v0-pubsub-pub && https://docs.ipfs.io/reference/http/api/#api-v0-pubsub-sub (Edit: This actually doesn't effect `sub`, details below!) Basically, we used to accept regular strings for pubsub topics over the HTTP API, now we require Base64URL encoding **prefixed with a 'u'** (Edit: As outlined below, this isn't *entirely* true, Base64URL is the default/recommended encoding, but technically any multibase should work https://github.com/multiformats/multibase, which is where the `u` comes from). Noting this down in an issue so I don't forget, I spun my wheels on this for a while not realizing this change happened, this is probably a good first issue though. Also not sure how other APIs have been effected. I believe the CLI handles this for you for example.
True
Document PubSub changes from go-ipfs v0.11 - This describes the issue very well: https://github.com/ipfs/go-ipfs/issues/8612 Caused by: https://github.com/ipfs/go-ipfs/issues/8343 Relevant areas: https://docs.ipfs.io/reference/http/api/#api-v0-pubsub-pub && https://docs.ipfs.io/reference/http/api/#api-v0-pubsub-sub (Edit: This actually doesn't effect `sub`, details below!) Basically, we used to accept regular strings for pubsub topics over the HTTP API, now we require Base64URL encoding **prefixed with a 'u'** (Edit: As outlined below, this isn't *entirely* true, Base64URL is the default/recommended encoding, but technically any multibase should work https://github.com/multiformats/multibase, which is where the `u` comes from). Noting this down in an issue so I don't forget, I spun my wheels on this for a while not realizing this change happened, this is probably a good first issue though. Also not sure how other APIs have been effected. I believe the CLI handles this for you for example.
main
document pubsub changes from go ipfs this describes the issue very well caused by relevant areas edit this actually doesn t effect sub details below basically we used to accept regular strings for pubsub topics over the http api now we require encoding prefixed with a u edit as outlined below this isn t entirely true is the default recommended encoding but technically any multibase should work which is where the u comes from noting this down in an issue so i don t forget i spun my wheels on this for a while not realizing this change happened this is probably a good first issue though also not sure how other apis have been effected i believe the cli handles this for you for example
1
7,164
2,885,911,024
IssuesEvent
2015-06-12 02:36:56
softlayer/sl-ember-components
https://api.github.com/repos/softlayer/sl-ember-components
closed
Add tests for sl-pagination-controls component
0 - Backlog sl-pagination-controls tests
<!--- @huboard:{"order":7.152557373046875e-07,"milestone_order":223,"custom_state":""} -->
1.0
Add tests for sl-pagination-controls component - <!--- @huboard:{"order":7.152557373046875e-07,"milestone_order":223,"custom_state":""} -->
non_main
add tests for sl pagination controls component huboard order milestone order custom state
0
3,132
12,017,890,005
IssuesEvent
2020-04-10 19:28:13
plotly/dash-docs
https://api.github.com/repos/plotly/dash-docs
opened
Reactivate devtools validation
dash-attribute-maintainability dash-type-bug
As part of https://github.com/plotly/dash-docs/pull/831 and https://github.com/plotly/dash-docs/pull/849, the DevTools asserts have been deactivated because the code wildcards implementation and callbacks validation uncovered pre-existing bugs. As the behavior/features all worked fine after the `Dash v1.11.0` release it was decided to go forward with the docs release and backfill.
True
Reactivate devtools validation - As part of https://github.com/plotly/dash-docs/pull/831 and https://github.com/plotly/dash-docs/pull/849, the DevTools asserts have been deactivated because the code wildcards implementation and callbacks validation uncovered pre-existing bugs. As the behavior/features all worked fine after the `Dash v1.11.0` release it was decided to go forward with the docs release and backfill.
main
reactivate devtools validation as part of and the devtools asserts have been deactivated because the code wildcards implementation and callbacks validation uncovered pre existing bugs as the behavior features all worked fine after the dash release it was decided to go forward with the docs release and backfill
1
816,439
30,599,421,934
IssuesEvent
2023-07-22 07:00:25
Weiver-project/Weiver
https://api.github.com/repos/Weiver-project/Weiver
closed
FE_[Feat]: 커뮤니티-검색 페이지
✨feat 🟡 Priority: Medium
## 📃To do List - [x] 뒤로가기 아이콘 클릭 시, 이전 페이지로 이동 - [x] 검색어 입력 후 해당 키워드와 매치하는 제목, 글 내용을 가진 Board 리스트 출력 - [x] 검색 결과 리스트[이미지, 글 제목, 내용]로 출력
1.0
FE_[Feat]: 커뮤니티-검색 페이지 - ## 📃To do List - [x] 뒤로가기 아이콘 클릭 시, 이전 페이지로 이동 - [x] 검색어 입력 후 해당 키워드와 매치하는 제목, 글 내용을 가진 Board 리스트 출력 - [x] 검색 결과 리스트[이미지, 글 제목, 내용]로 출력
non_main
fe 커뮤니티 검색 페이지 📃to do list 뒤로가기 아이콘 클릭 시 이전 페이지로 이동 검색어 입력 후 해당 키워드와 매치하는 제목 글 내용을 가진 board 리스트 출력 검색 결과 리스트 로 출력
0
193,722
22,216,296,967
IssuesEvent
2022-06-08 02:15:47
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2020-25670 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2020-25670 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in Linux Kernel where refcount leak in llcp_sock_bind() causing use-after-free which might lead to privilege escalations. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25670>CVE-2020-25670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-25670">https://nvd.nist.gov/vuln/detail/CVE-2020-25670</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: linux-libc-headers - 5.13;linux-yocto - 4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68,5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-25670 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2020-25670 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/llcp_sock.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in Linux Kernel where refcount leak in llcp_sock_bind() causing use-after-free which might lead to privilege escalations. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25670>CVE-2020-25670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-25670">https://nvd.nist.gov/vuln/detail/CVE-2020-25670</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: linux-libc-headers - 5.13;linux-yocto - 4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68,5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net nfc llcp sock c net nfc llcp sock c vulnerability details a vulnerability was found in linux kernel where refcount leak in llcp sock bind causing use after free which might lead to privilege escalations publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux libc headers linux yocto gitautoinc gitautoinc step up your open source security game with whitesource
0
689
4,234,990,544
IssuesEvent
2016-07-05 13:55:20
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
IA testing: combine fathead and longtail into api_result
Maintainer Input Requested
To have fathead and longtail PRs show up we need to also combine them into the `api_result`. https://github.com/duckduckgo/zeroclickinfo-spice/blob/master/share/spice/experimentalia/experimentalia.js#L9 ------ IA Page: http://duck.co/ia/view/experimentalia [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @vikashvverma
True
IA testing: combine fathead and longtail into api_result - To have fathead and longtail PRs show up we need to also combine them into the `api_result`. https://github.com/duckduckgo/zeroclickinfo-spice/blob/master/share/spice/experimentalia/experimentalia.js#L9 ------ IA Page: http://duck.co/ia/view/experimentalia [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @vikashvverma
main
ia testing combine fathead and longtail into api result to have fathead and longtail prs show up we need to also combine them into the api result ia page vikashvverma
1
4,357
22,038,100,971
IssuesEvent
2022-05-28 23:11:27
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[BUG] Copy to Clipboard doesn't work
🐛 Bug 👤 Awaiting Maintainer Response
### Environment Self-Hosted (Docker) ### Version 2.0.9 ### Describe the problem Export config to clipboard doesn't work (button "Copy to Clipboard"). Message in the console: ``` vue.runtime.esm.js:1897 TypeError: Cannot read properties of undefined (reading 'writeText') at r.copyConfigToClipboard (ExportConfigMenu.vue:83:27) at click (Button.vue?df23:1:384) at ne (vue.runtime.esm.js:1863:26) at HTMLButtonElement.i (vue.runtime.esm.js:2188:14) at HTMLButtonElement.Qs.o._wrapper (vue.runtime.esm.js:6961:25) ``` ### Additional info _No response_ ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [issue hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
True
[BUG] Copy to Clipboard doesn't work - ### Environment Self-Hosted (Docker) ### Version 2.0.9 ### Describe the problem Export config to clipboard doesn't work (button "Copy to Clipboard"). Message in the console: ``` vue.runtime.esm.js:1897 TypeError: Cannot read properties of undefined (reading 'writeText') at r.copyConfigToClipboard (ExportConfigMenu.vue:83:27) at click (Button.vue?df23:1:384) at ne (vue.runtime.esm.js:1863:26) at HTMLButtonElement.i (vue.runtime.esm.js:2188:14) at HTMLButtonElement.Qs.o._wrapper (vue.runtime.esm.js:6961:25) ``` ### Additional info _No response_ ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [issue hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
main
copy to clipboard doesn t work environment self hosted docker version describe the problem export config to clipboard doesn t work button copy to clipboard message in the console vue runtime esm js typeerror cannot read properties of undefined reading writetext at r copyconfigtoclipboard exportconfigmenu vue at click button vue at ne vue runtime esm js at htmlbuttonelement i vue runtime esm js at htmlbuttonelement qs o wrapper vue runtime esm js additional info no response please tick the boxes you are using a version of dashy check the first two digits of the version number you ve checked that this you ve checked the and guide you agree to the
1
143,307
11,542,230,010
IssuesEvent
2020-02-18 06:50:22
mattbearman/lime
https://api.github.com/repos/mattbearman/lime
opened
test
test
## [View in Saber Control Panel](https://app.saberfeedback.com/websites/a53d1139bceee3b3730c237aff8217d824ad1ccc/feedback/6e4e5c481c5c211ee316d9c4ca8c4af9e58887f0) ## ## Details ## **Star rating:** 3 **Submitted:** February 18, 2020 06:50 **Website:** Saber Website? **URL:** https://www.staging.saberfeedback.com/ **Operating System:** Mac OS X Catalina **Browser:** Safari 13.0.4 **Browser Size:** 1280 x 1354 **User Agent:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_2) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.0.4 Safari/605.1.15 ## Screenshot ## [View Screenshot](https://app.saberfeedback.com/websites/a53d1139bceee3b3730c237aff8217d824ad1ccc/feedback/6e4e5c481c5c211ee316d9c4ca8c4af9e58887f0) ## Browser Plugins ## - WebKit built-in PDF
1.0
test - ## [View in Saber Control Panel](https://app.saberfeedback.com/websites/a53d1139bceee3b3730c237aff8217d824ad1ccc/feedback/6e4e5c481c5c211ee316d9c4ca8c4af9e58887f0) ## ## Details ## **Star rating:** 3 **Submitted:** February 18, 2020 06:50 **Website:** Saber Website? **URL:** https://www.staging.saberfeedback.com/ **Operating System:** Mac OS X Catalina **Browser:** Safari 13.0.4 **Browser Size:** 1280 x 1354 **User Agent:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_2) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.0.4 Safari/605.1.15 ## Screenshot ## [View Screenshot](https://app.saberfeedback.com/websites/a53d1139bceee3b3730c237aff8217d824ad1ccc/feedback/6e4e5c481c5c211ee316d9c4ca8c4af9e58887f0) ## Browser Plugins ## - WebKit built-in PDF
non_main
test details star rating submitted february website saber website url operating system mac os x catalina browser safari browser size x user agent mozilla macintosh intel mac os x applewebkit khtml like gecko version safari screenshot browser plugins webkit built in pdf
0
182,250
14,908,442,014
IssuesEvent
2021-01-22 05:54:16
IBM/compliance-trestle
https://api.github.com/repos/IBM/compliance-trestle
closed
Create compliance-trestle website on github.io or similar
documentation enhancement help wanted
## Issue description / feature objectives For publicity it would be great to build a website for trestle. I already own https://github.com/compliance-trestle which would allow us to create a github page at compliance-trestle.github.io For this a decision needs to be made to 1) Type of website engine / template 2) Is it a simple website with some top level information? or a full documentation website.
1.0
Create compliance-trestle website on github.io or similar - ## Issue description / feature objectives For publicity it would be great to build a website for trestle. I already own https://github.com/compliance-trestle which would allow us to create a github page at compliance-trestle.github.io For this a decision needs to be made to 1) Type of website engine / template 2) Is it a simple website with some top level information? or a full documentation website.
non_main
create compliance trestle website on github io or similar issue description feature objectives for publicity it would be great to build a website for trestle i already own which would allow us to create a github page at compliance trestle github io for this a decision needs to be made to type of website engine template is it a simple website with some top level information or a full documentation website
0
74,355
20,144,542,393
IssuesEvent
2022-02-09 05:20:18
envoyproxy/envoy
https://api.github.com/repos/envoyproxy/envoy
closed
feature request: better build @com_googlesource_chromium_v8:build
area/build help wanted area/wasm
This target is internal spawn lots of tasks within. However, this target is considered as a single-core task from the view of bazel scheduler. Generally speaking, when you run `bazel build //source/exe:envoy` with N cores, N tasks will be scheduled and at most N clang is running. However, if the running task contains @com_googlesource_chromium_v8:build, ~2N clang is running, and the peak ram usage could be huge. A workaround: you may want to run `bazel build @com_googlesource_chromium_v8//:build` first, exhausting the cpu and ram. When the above task is completed, you can run `bazel build YOUR_REAL_TARGET` using the build cache of the above v8 build.
1.0
feature request: better build @com_googlesource_chromium_v8:build - This target is internal spawn lots of tasks within. However, this target is considered as a single-core task from the view of bazel scheduler. Generally speaking, when you run `bazel build //source/exe:envoy` with N cores, N tasks will be scheduled and at most N clang is running. However, if the running task contains @com_googlesource_chromium_v8:build, ~2N clang is running, and the peak ram usage could be huge. A workaround: you may want to run `bazel build @com_googlesource_chromium_v8//:build` first, exhausting the cpu and ram. When the above task is completed, you can run `bazel build YOUR_REAL_TARGET` using the build cache of the above v8 build.
non_main
feature request better build com googlesource chromium build this target is internal spawn lots of tasks within however this target is considered as a single core task from the view of bazel scheduler generally speaking when you run bazel build source exe envoy with n cores n tasks will be scheduled and at most n clang is running however if the running task contains com googlesource chromium build clang is running and the peak ram usage could be huge a workaround you may want to run bazel build com googlesource chromium build first exhausting the cpu and ram when the above task is completed you can run bazel build your real target using the build cache of the above build
0
4,002
18,674,088,411
IssuesEvent
2021-10-31 08:44:51
TWI-ArcOS/ArcOS-Environment
https://api.github.com/repos/TWI-ArcOS/ArcOS-Environment
opened
Refactoring ArcOS & implementing build systems
maintainability
As the codebase grows, it might be time to rethink the project structure, code and build systems. Right now, ArcOS doesn't use build systems aside from `electron-packager`, has no project-wide linter, is written in plain JavaScript and CSS without preprocessors, does not follow directory structure conventions and the codebase is – overall – very messy. It's also tightly coupled to the Electron runtime which makes it harder to run it using, say, the Tauri framework. The following list proposes a refactoring roadmap in chronological order: - [ ] Implement ESLint - [ ] Rethink the project structure - [ ] Implement ESBuild - [ ] Use ES Modules - [ ] Switch from CSS to SCSS - [ ] Switch from JavaScript to TypeScript - [ ] Improve classes to follow SOLID principles - [ ] Refactor the codebase This list will be updated from time to time. Feel free to add on to or to correct this list.
True
Refactoring ArcOS & implementing build systems - As the codebase grows, it might be time to rethink the project structure, code and build systems. Right now, ArcOS doesn't use build systems aside from `electron-packager`, has no project-wide linter, is written in plain JavaScript and CSS without preprocessors, does not follow directory structure conventions and the codebase is – overall – very messy. It's also tightly coupled to the Electron runtime which makes it harder to run it using, say, the Tauri framework. The following list proposes a refactoring roadmap in chronological order: - [ ] Implement ESLint - [ ] Rethink the project structure - [ ] Implement ESBuild - [ ] Use ES Modules - [ ] Switch from CSS to SCSS - [ ] Switch from JavaScript to TypeScript - [ ] Improve classes to follow SOLID principles - [ ] Refactor the codebase This list will be updated from time to time. Feel free to add on to or to correct this list.
main
refactoring arcos implementing build systems as the codebase grows it might be time to rethink the project structure code and build systems right now arcos doesn t use build systems aside from electron packager has no project wide linter is written in plain javascript and css without preprocessors does not follow directory structure conventions and the codebase is – overall – very messy it s also tightly coupled to the electron runtime which makes it harder to run it using say the tauri framework the following list proposes a refactoring roadmap in chronological order implement eslint rethink the project structure implement esbuild use es modules switch from css to scss switch from javascript to typescript improve classes to follow solid principles refactor the codebase this list will be updated from time to time feel free to add on to or to correct this list
1
18,677
6,628,926,894
IssuesEvent
2017-09-24 01:13:01
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
node_linux_dbg_uv_default_electron1_6 test are failing (in Linux Portability Master)
infra/BUILDPONY lang/node
https://grpc-testing.appspot.com/job/gRPC_portability_master_linux/1551/testReport/(root)/aggregate_tests/run_tests_node_linux_dbg_uv_default_electron1_6/ (it's passing on PRs because those run with --build_only). Possibly a missing dependency? ``` 2017-04-27 07:34:59,836 START: tools/run_tests/helper_scripts/pre_build_node_electron.sh 2017-04-27 07:35:16,827 PASSED: tools/run_tests/helper_scripts/pre_build_node_electron.sh [time=17.0sec; retries=0:0] 2017-04-27 07:35:16,828 START: tools/run_tests/helper_scripts/build_node_electron.sh 2017-04-27 07:35:47,844 PASSED: tools/run_tests/helper_scripts/build_node_electron.sh [time=31.0sec; retries=0:0] 2017-04-27 07:35:47,853 failed to detect port server Traceback (most recent call last): File "/var/local/git/grpc/tools/run_tests/python_utils/start_port_server.py", line 55, in start_port_server timeout=10).read()) File "/usr/lib/python2.7/urllib2.py", line 154, in urlopen return opener.open(url, data, timeout) File "/usr/lib/python2.7/urllib2.py", line 431, in open response = self._open(req, data) File "/usr/lib/python2.7/urllib2.py", line 449, in _open '_open', req) File "/usr/lib/python2.7/urllib2.py", line 409, in _call_chain result = func(*args) File "/usr/lib/python2.7/urllib2.py", line 1227, in http_open return self.do_open(httplib.HTTPConnection, req) File "/usr/lib/python2.7/urllib2.py", line 1197, in do_open raise URLError(err) URLError: <urlopen error [Errno 111] Connection refused> 2017-04-27 07:35:47,854 starting port_server, with log file /tmp/tmpE949w_ 2017-04-27 07:35:48,878 port server is up and ready 2017-04-27 07:35:48,878 START: tools/run_tests/helper_scripts/run_node_electron.sh 2017-04-27 07:35:49,297 N/A: version "v6" is not yet installed ++ dirname tools/run_tests/helper_scripts/run_node_electron.sh + cd tools/run_tests/helper_scripts/../../.. + test_directory=src/node/test + timeout=8000 + JUNIT_REPORT_PATH=src/node/report.xml + JUNIT_REPORT_STACK=1 + ./node_modules/.bin/xvfb-maybe ./node_modules/.bin/electron-mocha --timeout 8000 --reporter mocha-jenkins-reporter src/node/test /usr/bin/env: node : No such file or directory 2017-04-27 07:35:49,297 FAILED: tools/run_tests/helper_scripts/run_node_electron.sh [ret=127, pid=15113] ```
1.0
node_linux_dbg_uv_default_electron1_6 test are failing (in Linux Portability Master) - https://grpc-testing.appspot.com/job/gRPC_portability_master_linux/1551/testReport/(root)/aggregate_tests/run_tests_node_linux_dbg_uv_default_electron1_6/ (it's passing on PRs because those run with --build_only). Possibly a missing dependency? ``` 2017-04-27 07:34:59,836 START: tools/run_tests/helper_scripts/pre_build_node_electron.sh 2017-04-27 07:35:16,827 PASSED: tools/run_tests/helper_scripts/pre_build_node_electron.sh [time=17.0sec; retries=0:0] 2017-04-27 07:35:16,828 START: tools/run_tests/helper_scripts/build_node_electron.sh 2017-04-27 07:35:47,844 PASSED: tools/run_tests/helper_scripts/build_node_electron.sh [time=31.0sec; retries=0:0] 2017-04-27 07:35:47,853 failed to detect port server Traceback (most recent call last): File "/var/local/git/grpc/tools/run_tests/python_utils/start_port_server.py", line 55, in start_port_server timeout=10).read()) File "/usr/lib/python2.7/urllib2.py", line 154, in urlopen return opener.open(url, data, timeout) File "/usr/lib/python2.7/urllib2.py", line 431, in open response = self._open(req, data) File "/usr/lib/python2.7/urllib2.py", line 449, in _open '_open', req) File "/usr/lib/python2.7/urllib2.py", line 409, in _call_chain result = func(*args) File "/usr/lib/python2.7/urllib2.py", line 1227, in http_open return self.do_open(httplib.HTTPConnection, req) File "/usr/lib/python2.7/urllib2.py", line 1197, in do_open raise URLError(err) URLError: <urlopen error [Errno 111] Connection refused> 2017-04-27 07:35:47,854 starting port_server, with log file /tmp/tmpE949w_ 2017-04-27 07:35:48,878 port server is up and ready 2017-04-27 07:35:48,878 START: tools/run_tests/helper_scripts/run_node_electron.sh 2017-04-27 07:35:49,297 N/A: version "v6" is not yet installed ++ dirname tools/run_tests/helper_scripts/run_node_electron.sh + cd tools/run_tests/helper_scripts/../../.. + test_directory=src/node/test + timeout=8000 + JUNIT_REPORT_PATH=src/node/report.xml + JUNIT_REPORT_STACK=1 + ./node_modules/.bin/xvfb-maybe ./node_modules/.bin/electron-mocha --timeout 8000 --reporter mocha-jenkins-reporter src/node/test /usr/bin/env: node : No such file or directory 2017-04-27 07:35:49,297 FAILED: tools/run_tests/helper_scripts/run_node_electron.sh [ret=127, pid=15113] ```
non_main
node linux dbg uv default test are failing in linux portability master it s passing on prs because those run with build only possibly a missing dependency start tools run tests helper scripts pre build node electron sh passed tools run tests helper scripts pre build node electron sh start tools run tests helper scripts build node electron sh passed tools run tests helper scripts build node electron sh failed to detect port server traceback most recent call last file var local git grpc tools run tests python utils start port server py line in start port server timeout read file usr lib py line in urlopen return opener open url data timeout file usr lib py line in open response self open req data file usr lib py line in open open req file usr lib py line in call chain result func args file usr lib py line in http open return self do open httplib httpconnection req file usr lib py line in do open raise urlerror err urlerror starting port server with log file tmp port server is up and ready start tools run tests helper scripts run node electron sh n a version is not yet installed dirname tools run tests helper scripts run node electron sh cd tools run tests helper scripts test directory src node test timeout junit report path src node report xml junit report stack node modules bin xvfb maybe node modules bin electron mocha timeout reporter mocha jenkins reporter src node test usr bin env node no such file or directory failed tools run tests helper scripts run node electron sh
0
1,578
6,572,341,727
IssuesEvent
2017-09-11 01:32:52
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
deploy_helper crashes when unfinished_filename is a directory rather than a file
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME deploy_helper ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /home/mg/src/new-project/deployment/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT Linux (Ubuntu 16.04 LTS on the controller, Debian 8 on target) ##### SUMMARY When you create a directory named "{{ deploy_helper.unfinished_filename }}", deploy helper gets terribly confused when it tries to remove it: ##### STEPS TO REPRODUCE ``` - hosts: all tasks: - name: initialize deployment deploy_helper: path=/opt/app/ - name: mkdir /opt/app/* file: dest={{ deploy_helper.new_release_path }}/{{ item }} state=directory with_items: - "{{ deploy_helper.unfinished_filename }}" - frontend - backend # ... further build steps irrelevant for this bug ... - name: finalize the deployment deploy_helper: path: "{{ deploy_helper.project_path }}" release: "{{ deploy_helper.new_release }}" state: finalize ``` ##### EXPECTED RESULTS no crash ##### ACTUAL RESULTS crash: ``` fatal: [webserver]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_xGKPNM/ansible_module_deploy_helper.py\", line 475, in <module>\r\n main()\r\n File \"/tmp/ansible_xGKPNM/ansible_module_deploy_helper.py\", line 446, in main\r\n changes += deploy_helper.remove_unfinished_file(facts['new_release_path'])\r\n File \"/tmp/ansible_xGKPNM/ansible_module_deploy_helper.py\", line 341, in remove_unfinished_file\r\n os.remove(unfinished_file_path)\r\nOSError: [Errno 21] Aplankas: '/opt/app/releases/20160810131940/DEPLOY_UNFINISHED'\r\n", "msg": "MODULE FAILURE", "parsed": false} ```
True
deploy_helper crashes when unfinished_filename is a directory rather than a file - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME deploy_helper ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /home/mg/src/new-project/deployment/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT Linux (Ubuntu 16.04 LTS on the controller, Debian 8 on target) ##### SUMMARY When you create a directory named "{{ deploy_helper.unfinished_filename }}", deploy helper gets terribly confused when it tries to remove it: ##### STEPS TO REPRODUCE ``` - hosts: all tasks: - name: initialize deployment deploy_helper: path=/opt/app/ - name: mkdir /opt/app/* file: dest={{ deploy_helper.new_release_path }}/{{ item }} state=directory with_items: - "{{ deploy_helper.unfinished_filename }}" - frontend - backend # ... further build steps irrelevant for this bug ... - name: finalize the deployment deploy_helper: path: "{{ deploy_helper.project_path }}" release: "{{ deploy_helper.new_release }}" state: finalize ``` ##### EXPECTED RESULTS no crash ##### ACTUAL RESULTS crash: ``` fatal: [webserver]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_xGKPNM/ansible_module_deploy_helper.py\", line 475, in <module>\r\n main()\r\n File \"/tmp/ansible_xGKPNM/ansible_module_deploy_helper.py\", line 446, in main\r\n changes += deploy_helper.remove_unfinished_file(facts['new_release_path'])\r\n File \"/tmp/ansible_xGKPNM/ansible_module_deploy_helper.py\", line 341, in remove_unfinished_file\r\n os.remove(unfinished_file_path)\r\nOSError: [Errno 21] Aplankas: '/opt/app/releases/20160810131940/DEPLOY_UNFINISHED'\r\n", "msg": "MODULE FAILURE", "parsed": false} ```
main
deploy helper crashes when unfinished filename is a directory rather than a file issue type bug report component name deploy helper ansible version ansible config file home mg src new project deployment ansible cfg configured module search path default w o overrides os environment linux ubuntu lts on the controller debian on target summary when you create a directory named deploy helper unfinished filename deploy helper gets terribly confused when it tries to remove it steps to reproduce hosts all tasks name initialize deployment deploy helper path opt app name mkdir opt app file dest deploy helper new release path item state directory with items deploy helper unfinished filename frontend backend further build steps irrelevant for this bug name finalize the deployment deploy helper path deploy helper project path release deploy helper new release state finalize expected results no crash actual results crash fatal failed changed false failed true module stderr module stdout traceback most recent call last r n file tmp ansible xgkpnm ansible module deploy helper py line in r n main r n file tmp ansible xgkpnm ansible module deploy helper py line in main r n changes deploy helper remove unfinished file facts r n file tmp ansible xgkpnm ansible module deploy helper py line in remove unfinished file r n os remove unfinished file path r noserror aplankas opt app releases deploy unfinished r n msg module failure parsed false
1
4,452
23,160,174,199
IssuesEvent
2022-07-29 16:49:15
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[BUG] Unable to install docker image
🐛 Bug 👤 Awaiting Maintainer Response
### Environment Self-Hosted (Docker) ### System Unraid 6.10.3, Edge Version 103.0.1264.49 (Official build) (64-bit) ### Version latest ### Describe the problem I had Dashy running fine initially. However, I made so many changes trying to get a "feel" for it, that I messed it up beyond repair. I decided to just delete the docker image and re-install to start over. Now, I can not install Dashy either using the UnRAID Community App or from the cli with the command line. Li have cleared my entire docker environment thinking I might have a conflict: # Stop all containers docker stop `docker ps -qa` # Remove all containers docker rm `docker ps -qa` # Remove all images docker rmi -f `docker images -qa ` # Remove all volumes docker volume rm $(docker volume ls -qf) # Remove all networks docker network rm `docker network ls -q` # Your installation should now be all fresh and clean. # The following commands should not output any items: docker ps -a docker images -a docker volume ls # The following command show only show the default networks: docker network ls But when trying to install via the cli I get this error: docker run -d -p 8080:80 -v ~/my-conf.yml:/app/public/conf.yml --name Dashy --restart=always lissy93/dashy:latest Unable to find image 'lissy93/dashy:latest' locally latest: Pulling from lissy93/dashy 59bf1c3509f3: Pull complete b616ac4a64bf: Pull complete 3b9e1e8ab9ce: Pull complete 3507ddbf3909: Pull complete 88ffe1153cc4: Pull complete 7b23c8336106: Pull complete 91ae094d5121: Pull complete 3aa391065cb8: Pull complete Digest: sha256:bc59cc6029b52dc93afbf94c37deb867d54edd6cc8e19bb22eecf6e805b88942 Status: Downloaded newer image for lissy93/dashy:latest ab2d9a73c95c546f24ed38b4e05534ac25f3e641941ea2d1707482810d6bbd9f docker: Error response from daemon: failed to create shim: OCI runtime create failed: container_linux.go:380: starting container process caused: process_linux.go:545: container init caused: rootfs_linux.go:75: mounting "/root/my-conf.yml" to rootfs at "/app/public/conf.yml" caused: mount through procfd: not a directory: unknown: Are you trying to mount a directory onto a file (or vice-versa)? Check if the specified host path exists and is the expected type. If I wipe the docker environment again and try to install via the UnRAID Community App, I get this error: root@localhost:# /usr/local/emhttp/plugins/dynamix.docker.manager/scripts/docker run -d --name='dashy' --net='bridge' -e TZ="America/Los_Angeles" -e HOST_OS="Unraid" -e HOST_HOSTNAME="Unraid-1" -e HOST_CONTAINERNAME="dashy" -e 'UID'='99' -e 'GID'='100' -l net.unraid.docker.managed=dockerman -l net.unraid.docker.webui='http://[IP]:[PORT:80]/' -l net.unraid.docker.icon='https://raw.githubusercontent.com/A75G/docker-templates/master/templates/icons/dashy.png' -p '4000:80/tcp' -v '/mnt/user/appdata/dashy/conf.yml':'/app/public/conf.yml':'rw' -v '/mnt/user/appdata/dashy/icons/':'/app/public/item-icons/':'rw' 'lissy93/dashy:latest' b394fd10478bbbd2321c1d5b5347eca6480729bd43a7feba15fb416a41fe0566 docker: Error response from daemon: failed to create shim: OCI runtime create failed: container_linux.go:380: starting container process caused: process_linux.go:545: container init caused: rootfs_linux.go:75: mounting "/mnt/user/appdata/dashy/conf.yml" to rootfs at "/app/public/conf.yml" caused: mount through procfd: not a directory: unknown: Are you trying to mount a directory onto a file (or vice-versa)? Check if the specified host path exists and is the expected type. The command failed. Any help getting this resolved would be greatly appreciated. ### Additional info _No response_ ### Please tick the boxes - [X] You have explained the issue clearly, and included all relevant info - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy - [X] You've checked that this [issue hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide ![important](https://img.shields.io/badge/Important!-F6094E) - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
True
[BUG] Unable to install docker image - ### Environment Self-Hosted (Docker) ### System Unraid 6.10.3, Edge Version 103.0.1264.49 (Official build) (64-bit) ### Version latest ### Describe the problem I had Dashy running fine initially. However, I made so many changes trying to get a "feel" for it, that I messed it up beyond repair. I decided to just delete the docker image and re-install to start over. Now, I can not install Dashy either using the UnRAID Community App or from the cli with the command line. Li have cleared my entire docker environment thinking I might have a conflict: # Stop all containers docker stop `docker ps -qa` # Remove all containers docker rm `docker ps -qa` # Remove all images docker rmi -f `docker images -qa ` # Remove all volumes docker volume rm $(docker volume ls -qf) # Remove all networks docker network rm `docker network ls -q` # Your installation should now be all fresh and clean. # The following commands should not output any items: docker ps -a docker images -a docker volume ls # The following command show only show the default networks: docker network ls But when trying to install via the cli I get this error: docker run -d -p 8080:80 -v ~/my-conf.yml:/app/public/conf.yml --name Dashy --restart=always lissy93/dashy:latest Unable to find image 'lissy93/dashy:latest' locally latest: Pulling from lissy93/dashy 59bf1c3509f3: Pull complete b616ac4a64bf: Pull complete 3b9e1e8ab9ce: Pull complete 3507ddbf3909: Pull complete 88ffe1153cc4: Pull complete 7b23c8336106: Pull complete 91ae094d5121: Pull complete 3aa391065cb8: Pull complete Digest: sha256:bc59cc6029b52dc93afbf94c37deb867d54edd6cc8e19bb22eecf6e805b88942 Status: Downloaded newer image for lissy93/dashy:latest ab2d9a73c95c546f24ed38b4e05534ac25f3e641941ea2d1707482810d6bbd9f docker: Error response from daemon: failed to create shim: OCI runtime create failed: container_linux.go:380: starting container process caused: process_linux.go:545: container init caused: rootfs_linux.go:75: mounting "/root/my-conf.yml" to rootfs at "/app/public/conf.yml" caused: mount through procfd: not a directory: unknown: Are you trying to mount a directory onto a file (or vice-versa)? Check if the specified host path exists and is the expected type. If I wipe the docker environment again and try to install via the UnRAID Community App, I get this error: root@localhost:# /usr/local/emhttp/plugins/dynamix.docker.manager/scripts/docker run -d --name='dashy' --net='bridge' -e TZ="America/Los_Angeles" -e HOST_OS="Unraid" -e HOST_HOSTNAME="Unraid-1" -e HOST_CONTAINERNAME="dashy" -e 'UID'='99' -e 'GID'='100' -l net.unraid.docker.managed=dockerman -l net.unraid.docker.webui='http://[IP]:[PORT:80]/' -l net.unraid.docker.icon='https://raw.githubusercontent.com/A75G/docker-templates/master/templates/icons/dashy.png' -p '4000:80/tcp' -v '/mnt/user/appdata/dashy/conf.yml':'/app/public/conf.yml':'rw' -v '/mnt/user/appdata/dashy/icons/':'/app/public/item-icons/':'rw' 'lissy93/dashy:latest' b394fd10478bbbd2321c1d5b5347eca6480729bd43a7feba15fb416a41fe0566 docker: Error response from daemon: failed to create shim: OCI runtime create failed: container_linux.go:380: starting container process caused: process_linux.go:545: container init caused: rootfs_linux.go:75: mounting "/mnt/user/appdata/dashy/conf.yml" to rootfs at "/app/public/conf.yml" caused: mount through procfd: not a directory: unknown: Are you trying to mount a directory onto a file (or vice-versa)? Check if the specified host path exists and is the expected type. The command failed. Any help getting this resolved would be greatly appreciated. ### Additional info _No response_ ### Please tick the boxes - [X] You have explained the issue clearly, and included all relevant info - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy - [X] You've checked that this [issue hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide ![important](https://img.shields.io/badge/Important!-F6094E) - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
main
unable to install docker image environment self hosted docker system unraid edge version official build bit version latest describe the problem i had dashy running fine initially however i made so many changes trying to get a feel for it that i messed it up beyond repair i decided to just delete the docker image and re install to start over now i can not install dashy either using the unraid community app or from the cli with the command line li have cleared my entire docker environment thinking i might have a conflict stop all containers docker stop docker ps qa remove all containers docker rm docker ps qa remove all images docker rmi f docker images qa remove all volumes docker volume rm docker volume ls qf remove all networks docker network rm docker network ls q your installation should now be all fresh and clean the following commands should not output any items docker ps a docker images a docker volume ls the following command show only show the default networks docker network ls but when trying to install via the cli i get this error docker run d p v my conf yml app public conf yml name dashy restart always dashy latest unable to find image dashy latest locally latest pulling from dashy pull complete pull complete pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for dashy latest docker error response from daemon failed to create shim oci runtime create failed container linux go starting container process caused process linux go container init caused rootfs linux go mounting root my conf yml to rootfs at app public conf yml caused mount through procfd not a directory unknown are you trying to mount a directory onto a file or vice versa check if the specified host path exists and is the expected type if i wipe the docker environment again and try to install via the unraid community app i get this error root localhost usr local emhttp plugins dynamix docker manager scripts docker run d name dashy net bridge e tz america los angeles e host os unraid e host hostname unraid e host containername dashy e uid e gid l net unraid docker managed dockerman l net unraid docker webui http l net unraid docker icon p tcp v mnt user appdata dashy conf yml app public conf yml rw v mnt user appdata dashy icons app public item icons rw dashy latest docker error response from daemon failed to create shim oci runtime create failed container linux go starting container process caused process linux go container init caused rootfs linux go mounting mnt user appdata dashy conf yml to rootfs at app public conf yml caused mount through procfd not a directory unknown are you trying to mount a directory onto a file or vice versa check if the specified host path exists and is the expected type the command failed any help getting this resolved would be greatly appreciated additional info no response please tick the boxes you have explained the issue clearly and included all relevant info you are using a version of dashy you ve checked that this you ve checked the and guide you agree to the
1
4,219
20,870,054,712
IssuesEvent
2022-03-22 11:03:38
precice/precice
https://api.github.com/repos/precice/precice
opened
Remove `testBug`
maintainability
I found a test where the purpose is not clear to me: https://github.com/precice/precice/blob/285e85f6846b407e0fffaa136cf9438fb84be9ab/src/precice/tests/SerialTests.cpp#L1492-L1504 I did a bit of archeology, but did not really find anything helpful in the commit history, as well: https://github.com/precice/precice/blob/9bbd46812f97a68e0a1ed5fa10b97f416d49555c/src/precice/tests/couplingmode/SolverInterfaceTest.cpp#L1605 Does anybody know about the original purpose of this test? As far as I see it's doing something with subcycling. Maybe there is more to it and I don't see it. If the only purpose of this test is the testing of subcycling, I'm adding some new tests in #1186, which are testing this feature and should be more focused. Did anybody actually ever see this test failing? I would suggest to delete the test with the next release, if there are no new insights.
True
Remove `testBug` - I found a test where the purpose is not clear to me: https://github.com/precice/precice/blob/285e85f6846b407e0fffaa136cf9438fb84be9ab/src/precice/tests/SerialTests.cpp#L1492-L1504 I did a bit of archeology, but did not really find anything helpful in the commit history, as well: https://github.com/precice/precice/blob/9bbd46812f97a68e0a1ed5fa10b97f416d49555c/src/precice/tests/couplingmode/SolverInterfaceTest.cpp#L1605 Does anybody know about the original purpose of this test? As far as I see it's doing something with subcycling. Maybe there is more to it and I don't see it. If the only purpose of this test is the testing of subcycling, I'm adding some new tests in #1186, which are testing this feature and should be more focused. Did anybody actually ever see this test failing? I would suggest to delete the test with the next release, if there are no new insights.
main
remove testbug i found a test where the purpose is not clear to me i did a bit of archeology but did not really find anything helpful in the commit history as well does anybody know about the original purpose of this test as far as i see it s doing something with subcycling maybe there is more to it and i don t see it if the only purpose of this test is the testing of subcycling i m adding some new tests in which are testing this feature and should be more focused did anybody actually ever see this test failing i would suggest to delete the test with the next release if there are no new insights
1
2,370
8,475,814,608
IssuesEvent
2018-10-24 20:01:35
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
PagerDuty module uses decommissioned v1 PagerDuty API
affects_2.6 bug module needs_maintainer support:community
##### SUMMARY PagerDuty has decommissioned their v1 API endpoints. As a result, the pagerduty Ansible window which uses them no longer functions. https://v2.developer.pagerduty.com/docs/v1-rest-api-decommissioning-faq ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME pagerduty ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.6.4 ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> <!--- Paste example playbooks or commands between quotes below --> ```yaml ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes --> ```TASK [monitor/pagerduty : Create maintenance window] *************************** Tuesday 23 October 2018 18:31:40 +0000 (0:00:00.161) 0:00:01.617 ******* failed: [ansible-01.dev2 -> localhost] (item=XXXXXXX) => {"changed": false, "item": "PWVNHTL", "msg": "failed to create the window: HTTP Error 400: Bad Request"} ```
True
PagerDuty module uses decommissioned v1 PagerDuty API - ##### SUMMARY PagerDuty has decommissioned their v1 API endpoints. As a result, the pagerduty Ansible window which uses them no longer functions. https://v2.developer.pagerduty.com/docs/v1-rest-api-decommissioning-faq ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME pagerduty ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.6.4 ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> <!--- Paste example playbooks or commands between quotes below --> ```yaml ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes --> ```TASK [monitor/pagerduty : Create maintenance window] *************************** Tuesday 23 October 2018 18:31:40 +0000 (0:00:00.161) 0:00:01.617 ******* failed: [ansible-01.dev2 -> localhost] (item=XXXXXXX) => {"changed": false, "item": "PWVNHTL", "msg": "failed to create the window: HTTP Error 400: Bad Request"} ```
main
pagerduty module uses decommissioned pagerduty api summary pagerduty has decommissioned their api endpoints as a result the pagerduty ansible window which uses them no longer functions issue type bug report component name pagerduty ansible version paste below ansible configuration paste below os environment steps to reproduce yaml expected results actual results task tuesday october failed item xxxxxxx changed false item pwvnhtl msg failed to create the window http error bad request
1
2,483
8,639,918,711
IssuesEvent
2018-11-23 22:42:23
F5OEO/rpitx
https://api.github.com/repos/F5OEO/rpitx
closed
PSK with gnuradio and rpitx
V1 related (not maintained)
Hi, i'm trying to send a simple text file via PSK modulation by using gnuRadio and rpitx. I modulate with gnu radio and send via tcp socket to rpitx listening on tcp : ` nc -l 9999 | sudo rpitx -i- -m IQFLOAT -f 67000` The modulator : ![image](https://user-images.githubusercontent.com/44431383/47424364-cc1fe800-d787-11e8-984f-f01ccbfea409.png) And i demodulate via gnu-radio by using rtl sdr. ![image](https://user-images.githubusercontent.com/44431383/47424744-afd07b00-d788-11e8-8da7-748dfd36ea6d.png) I launch all these process and i revceive nothing,my ouptut file still at 0 bit , i can see on the spectrum and the waterfall that is a modulation but on the eye diagramm, i see that : ![capture d ecran du 2018-10-24 12-43-50](https://user-images.githubusercontent.com/44431383/47425793-b8768080-d78b-11e8-90ed-4ecd40f178f8.png) ![capture d ecran du 2018-10-24 12-44-34](https://user-images.githubusercontent.com/44431383/47425806-c2987f00-d78b-11e8-8507-5fa3b9584884.png) I try some different configuration about filtering and i get sometimes better constellation, but still transmit nothing. Sometimes, i have the "circle" of psk modulation but i clearly put in parameters 2 constellations point... It's for personnal testing, i'm just trying to have a simple and functionnal digital modulation. If i was wrong somewhere, if you have any idea ? I take any advice, thank you in advance.
True
PSK with gnuradio and rpitx - Hi, i'm trying to send a simple text file via PSK modulation by using gnuRadio and rpitx. I modulate with gnu radio and send via tcp socket to rpitx listening on tcp : ` nc -l 9999 | sudo rpitx -i- -m IQFLOAT -f 67000` The modulator : ![image](https://user-images.githubusercontent.com/44431383/47424364-cc1fe800-d787-11e8-984f-f01ccbfea409.png) And i demodulate via gnu-radio by using rtl sdr. ![image](https://user-images.githubusercontent.com/44431383/47424744-afd07b00-d788-11e8-8da7-748dfd36ea6d.png) I launch all these process and i revceive nothing,my ouptut file still at 0 bit , i can see on the spectrum and the waterfall that is a modulation but on the eye diagramm, i see that : ![capture d ecran du 2018-10-24 12-43-50](https://user-images.githubusercontent.com/44431383/47425793-b8768080-d78b-11e8-90ed-4ecd40f178f8.png) ![capture d ecran du 2018-10-24 12-44-34](https://user-images.githubusercontent.com/44431383/47425806-c2987f00-d78b-11e8-8507-5fa3b9584884.png) I try some different configuration about filtering and i get sometimes better constellation, but still transmit nothing. Sometimes, i have the "circle" of psk modulation but i clearly put in parameters 2 constellations point... It's for personnal testing, i'm just trying to have a simple and functionnal digital modulation. If i was wrong somewhere, if you have any idea ? I take any advice, thank you in advance.
main
psk with gnuradio and rpitx hi i m trying to send a simple text file via psk modulation by using gnuradio and rpitx i modulate with gnu radio and send via tcp socket to rpitx listening on tcp nc l sudo rpitx i m iqfloat f the modulator and i demodulate via gnu radio by using rtl sdr i launch all these process and i revceive nothing my ouptut file still at bit i can see on the spectrum and the waterfall that is a modulation but on the eye diagramm i see that i try some different configuration about filtering and i get sometimes better constellation but still transmit nothing sometimes i have the circle of psk modulation but i clearly put in parameters constellations point it s for personnal testing i m just trying to have a simple and functionnal digital modulation if i was wrong somewhere if you have any idea i take any advice thank you in advance
1
1,748
6,574,942,739
IssuesEvent
2017-09-11 14:34:03
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_image_facts does not search for tag
affects_2.1 bug_report cloud docker waiting_on_maintainer
<!--- Please do not report issues/requests related to Ansible modules here !! Report them to the appropriate modules-core or modules-extras project: - https://github.com/ansible/ansible-modules-core/issues - https://github.com/ansible/ansible-modules-extras/issues Also verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/task/feature --> docker_image_facts ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Ubuntu 14.04.5/Debian Jessie Docker Version: ``` Client: Version: 1.12.2 API version: 1.24 Go version: go1.6.3 Git commit: bb80604 Built: Tue Oct 11 17:43:41 2016 OS/Arch: linux/amd64 Server: Version: 1.12.2 API version: 1.24 Go version: go1.6.3 Git commit: bb80604 Built: Tue Oct 11 17:43:41 2016 OS/Arch: linux/amd64 ``` ##### SUMMARY <!--- Explain the problem briefly --> docker_image_facts retuirns all named images regardless of tag, even when specifying a tag in the name attribute of the module. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Specify tag in docker-image_facts name attribute using `:<tag-name>` Latest is returned in the image array, along with any other debian image in the local registry. <!--- Paste example playbooks or commands between quotes below --> ``` - hosts: localhost connection: local tasks: - docker_image_facts: name="debian:latest" - docker_image_facts: name="debian:not-latest" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The latest image not being returned in the second task. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [docker_image_facts] ****************************************************** PLAYBOOK: docker-test.yml ****************************************************** 1 plays in docker-test.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466 `" && echo ansible-tmp-1476812867.36-239285172461466="` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpLHbQfE TO /root/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466/setup <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466/setup; rm -rf "/root/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [docker_image_facts] ****************************************************** task path: /root/docker-test.yml:4 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847 `" && echo ansible-tmp-1476812867.85-152102133665847="` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpvxleDe TO /root/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847/docker_image_facts <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847/docker_image_facts; rm -rf "/root/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "images": [{"Architecture": "amd64", "Author": "", "Comment": "", "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/bash"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Container": "27a6f130fb20b127801c1179ff49cc6adcb252c6d91f5955421e4c66a72ced31", "ContainerConfig": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/sh", "-c", "#(nop) ", "CMD [\"/bin/bash\"]"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-09-23T18:08:51.133779867Z", "DockerVersion": "1.12.1", "GraphDriver": {"Data": {"DeviceId": "2", "DeviceName": "docker-202:2-403255-60c0ce068752b2eacce3a5c0985bd008bb4641443697174f5eab997d09c81ef6", "DeviceSize": "10737418240"}, "Name": "devicemapper"}, "Id": "sha256:ddf73f48a05d97e4f473d0b4ccb53383cbb0647d10e34b62d68bfc859cc6bcf9", "Os": "linux", "Parent": "", "RepoDigests": ["debian@sha256:677f184a5969847c0ad91d30cf1f0b925cd321e6c66e3ed5fbf9858f58425d1a"], "RepoTags": ["debian:latest"], "RootFS": {"Layers": ["sha256:142a601d97936307e75220c35dde0348971a9584c21e7cb42e1f7004005432ab"], "Type": "layers"}, "Size": 122988258, "VirtualSize": 122988258}], "invocation": {"module_args": {"api_version": null, "cacert_path": null, "cert_path": null, "debug": false, "docker_host": null, "filter_logger": false, "key_path": null, "name": ["debian:latest"], "ssl_version": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null}, "module_name": "docker_image_facts"}} TASK [docker_image_facts] ****************************************************** task path: /root/docker-test.yml:6 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637 `" && echo ansible-tmp-1476812868.02-253123017828637="` echo $HOME/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmp0oNjN1 TO /root/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637/docker_image_facts <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637/docker_image_facts; rm -rf "/root/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "images": [{"Architecture": "amd64", "Author": "", "Comment": "", "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/bash"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Container": "27a6f130fb20b127801c1179ff49cc6adcb252c6d91f5955421e4c66a72ced31", "ContainerConfig": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/sh", "-c", "#(nop) ", "CMD [\"/bin/bash\"]"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-09-23T18:08:51.133779867Z", "DockerVersion": "1.12.1", "GraphDriver": {"Data": {"DeviceId": "2", "DeviceName": "docker-202:2-403255-60c0ce068752b2eacce3a5c0985bd008bb4641443697174f5eab997d09c81ef6", "DeviceSize": "10737418240"}, "Name": "devicemapper"}, "Id": "sha256:ddf73f48a05d97e4f473d0b4ccb53383cbb0647d10e34b62d68bfc859cc6bcf9", "Os": "linux", "Parent": "", "RepoDigests": ["debian@sha256:677f184a5969847c0ad91d30cf1f0b925cd321e6c66e3ed5fbf9858f58425d1a"], "RepoTags": ["debian:latest"], "RootFS": {"Layers": ["sha256:142a601d97936307e75220c35dde0348971a9584c21e7cb42e1f7004005432ab"], "Type": "layers"}, "Size": 122988258, "VirtualSize": 122988258}], "invocation": {"module_args": {"api_version": null, "cacert_path": null, "cert_path": null, "debug": false, "docker_host": null, "filter_logger": false, "key_path": null, "name": ["debian:not-latest"], "ssl_version": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null}, "module_name": "docker_image_facts"}} PLAY RECAP ********************************************************************* localhost : ok=3 changed=0 unreachable=0 failed=0 ```
True
docker_image_facts does not search for tag - <!--- Please do not report issues/requests related to Ansible modules here !! Report them to the appropriate modules-core or modules-extras project: - https://github.com/ansible/ansible-modules-core/issues - https://github.com/ansible/ansible-modules-extras/issues Also verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/task/feature --> docker_image_facts ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Ubuntu 14.04.5/Debian Jessie Docker Version: ``` Client: Version: 1.12.2 API version: 1.24 Go version: go1.6.3 Git commit: bb80604 Built: Tue Oct 11 17:43:41 2016 OS/Arch: linux/amd64 Server: Version: 1.12.2 API version: 1.24 Go version: go1.6.3 Git commit: bb80604 Built: Tue Oct 11 17:43:41 2016 OS/Arch: linux/amd64 ``` ##### SUMMARY <!--- Explain the problem briefly --> docker_image_facts retuirns all named images regardless of tag, even when specifying a tag in the name attribute of the module. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Specify tag in docker-image_facts name attribute using `:<tag-name>` Latest is returned in the image array, along with any other debian image in the local registry. <!--- Paste example playbooks or commands between quotes below --> ``` - hosts: localhost connection: local tasks: - docker_image_facts: name="debian:latest" - docker_image_facts: name="debian:not-latest" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The latest image not being returned in the second task. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [docker_image_facts] ****************************************************** PLAYBOOK: docker-test.yml ****************************************************** 1 plays in docker-test.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466 `" && echo ansible-tmp-1476812867.36-239285172461466="` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpLHbQfE TO /root/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466/setup <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466/setup; rm -rf "/root/.ansible/tmp/ansible-tmp-1476812867.36-239285172461466/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [docker_image_facts] ****************************************************** task path: /root/docker-test.yml:4 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847 `" && echo ansible-tmp-1476812867.85-152102133665847="` echo $HOME/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpvxleDe TO /root/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847/docker_image_facts <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847/docker_image_facts; rm -rf "/root/.ansible/tmp/ansible-tmp-1476812867.85-152102133665847/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "images": [{"Architecture": "amd64", "Author": "", "Comment": "", "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/bash"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Container": "27a6f130fb20b127801c1179ff49cc6adcb252c6d91f5955421e4c66a72ced31", "ContainerConfig": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/sh", "-c", "#(nop) ", "CMD [\"/bin/bash\"]"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-09-23T18:08:51.133779867Z", "DockerVersion": "1.12.1", "GraphDriver": {"Data": {"DeviceId": "2", "DeviceName": "docker-202:2-403255-60c0ce068752b2eacce3a5c0985bd008bb4641443697174f5eab997d09c81ef6", "DeviceSize": "10737418240"}, "Name": "devicemapper"}, "Id": "sha256:ddf73f48a05d97e4f473d0b4ccb53383cbb0647d10e34b62d68bfc859cc6bcf9", "Os": "linux", "Parent": "", "RepoDigests": ["debian@sha256:677f184a5969847c0ad91d30cf1f0b925cd321e6c66e3ed5fbf9858f58425d1a"], "RepoTags": ["debian:latest"], "RootFS": {"Layers": ["sha256:142a601d97936307e75220c35dde0348971a9584c21e7cb42e1f7004005432ab"], "Type": "layers"}, "Size": 122988258, "VirtualSize": 122988258}], "invocation": {"module_args": {"api_version": null, "cacert_path": null, "cert_path": null, "debug": false, "docker_host": null, "filter_logger": false, "key_path": null, "name": ["debian:latest"], "ssl_version": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null}, "module_name": "docker_image_facts"}} TASK [docker_image_facts] ****************************************************** task path: /root/docker-test.yml:6 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637 `" && echo ansible-tmp-1476812868.02-253123017828637="` echo $HOME/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmp0oNjN1 TO /root/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637/docker_image_facts <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637/docker_image_facts; rm -rf "/root/.ansible/tmp/ansible-tmp-1476812868.02-253123017828637/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "images": [{"Architecture": "amd64", "Author": "", "Comment": "", "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/bash"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Container": "27a6f130fb20b127801c1179ff49cc6adcb252c6d91f5955421e4c66a72ced31", "ContainerConfig": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/sh", "-c", "#(nop) ", "CMD [\"/bin/bash\"]"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "383850eeb47b", "Image": "sha256:3681da375b325973f1297e0f64a2feef4fcae77715d54c34aee2188431ed6f46", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-09-23T18:08:51.133779867Z", "DockerVersion": "1.12.1", "GraphDriver": {"Data": {"DeviceId": "2", "DeviceName": "docker-202:2-403255-60c0ce068752b2eacce3a5c0985bd008bb4641443697174f5eab997d09c81ef6", "DeviceSize": "10737418240"}, "Name": "devicemapper"}, "Id": "sha256:ddf73f48a05d97e4f473d0b4ccb53383cbb0647d10e34b62d68bfc859cc6bcf9", "Os": "linux", "Parent": "", "RepoDigests": ["debian@sha256:677f184a5969847c0ad91d30cf1f0b925cd321e6c66e3ed5fbf9858f58425d1a"], "RepoTags": ["debian:latest"], "RootFS": {"Layers": ["sha256:142a601d97936307e75220c35dde0348971a9584c21e7cb42e1f7004005432ab"], "Type": "layers"}, "Size": 122988258, "VirtualSize": 122988258}], "invocation": {"module_args": {"api_version": null, "cacert_path": null, "cert_path": null, "debug": false, "docker_host": null, "filter_logger": false, "key_path": null, "name": ["debian:not-latest"], "ssl_version": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null}, "module_name": "docker_image_facts"}} PLAY RECAP ********************************************************************* localhost : ok=3 changed=0 unreachable=0 failed=0 ```
main
docker image facts does not search for tag please do not report issues requests related to ansible modules here report them to the appropriate modules core or modules extras project also verify first that your issue request is not already reported in github issue type bug report component name docker image facts ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific ubuntu debian jessie docker version client version api version go version git commit built tue oct os arch linux server version api version go version git commit built tue oct os arch linux summary docker image facts retuirns all named images regardless of tag even when specifying a tag in the name attribute of the module steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used specify tag in docker image facts name attribute using latest is returned in the image array along with any other debian image in the local registry hosts localhost connection local tasks docker image facts name debian latest docker image facts name debian not latest expected results the latest image not being returned in the second task actual results play task ok task playbook docker test yml plays in docker test yml play task establish local connection for user root exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmplhbqfe to root ansible tmp ansible tmp setup exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python root ansible tmp ansible tmp setup rm rf root ansible tmp ansible tmp dev null sleep ok task task path root docker test yml establish local connection for user root exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpvxlede to root ansible tmp ansible tmp docker image facts exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python root ansible tmp ansible tmp docker image facts rm rf root ansible tmp ansible tmp dev null sleep ok changed false images domainname entrypoint null env hostname image labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir container containerconfig attachstderr false attachstdin false attachstdout false cmd domainname entrypoint null env hostname image labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir created dockerversion graphdriver data deviceid devicename docker devicesize name devicemapper id os linux parent repodigests repotags rootfs layers type layers size virtualsize invocation module args api version null cacert path null cert path null debug false docker host null filter logger false key path null name ssl version null timeout null tls null tls hostname null tls verify null module name docker image facts task task path root docker test yml establish local connection for user root exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to root ansible tmp ansible tmp docker image facts exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python root ansible tmp ansible tmp docker image facts rm rf root ansible tmp ansible tmp dev null sleep ok changed false images domainname entrypoint null env hostname image labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir container containerconfig attachstderr false attachstdin false attachstdout false cmd domainname entrypoint null env hostname image labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir created dockerversion graphdriver data deviceid devicename docker devicesize name devicemapper id os linux parent repodigests repotags rootfs layers type layers size virtualsize invocation module args api version null cacert path null cert path null debug false docker host null filter logger false key path null name ssl version null timeout null tls null tls hostname null tls verify null module name docker image facts play recap localhost ok changed unreachable failed
1
5,333
26,920,257,451
IssuesEvent
2023-02-07 09:57:40
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Cleanup the multiple SLF4j bindings
maintainability maven
**Describe the bug** SLF4J API is designed to bind with one and only one underlying logging framework at a time. If more than one binding is present on the class path, SLF4J will emit a warning, listing the location of those bindings. **To Reproduce** Steps to reproduce the behavior: 1. Run ./refine **Current Results** ``` 22:40:43.060 [ refine_server] Creating new workspace directory /home/thad/.local/share/openrefine (383ms) SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/home/thad/OpenRefine/server/target/lib/slf4j-log4j12-1.7.18.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/home/thad/OpenRefine/main/webapp/WEB-INF/lib/slf4j-log4j12-1.7.18.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] 22:40:43.095 [ refine] Starting OpenRefine 3.1-beta [TRUNK]... (35ms) ``` **Expected behavior** No multiple SLF4J bindings warning. **Desktop (please complete the following information):** - OS: Ubuntu 18.04 - Browser Version: Firefox latest - JRE or JDK Version: OpenJDK 8 **OpenRefine (please complete the following information):** - Version: Trunk (master) **Additional Context** https://www.slf4j.org/codes.html#multiple_bindings Maybe we want to exclude as runtime dependencies in maven ? and only as compile time ?
True
Cleanup the multiple SLF4j bindings - **Describe the bug** SLF4J API is designed to bind with one and only one underlying logging framework at a time. If more than one binding is present on the class path, SLF4J will emit a warning, listing the location of those bindings. **To Reproduce** Steps to reproduce the behavior: 1. Run ./refine **Current Results** ``` 22:40:43.060 [ refine_server] Creating new workspace directory /home/thad/.local/share/openrefine (383ms) SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/home/thad/OpenRefine/server/target/lib/slf4j-log4j12-1.7.18.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/home/thad/OpenRefine/main/webapp/WEB-INF/lib/slf4j-log4j12-1.7.18.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] 22:40:43.095 [ refine] Starting OpenRefine 3.1-beta [TRUNK]... (35ms) ``` **Expected behavior** No multiple SLF4J bindings warning. **Desktop (please complete the following information):** - OS: Ubuntu 18.04 - Browser Version: Firefox latest - JRE or JDK Version: OpenJDK 8 **OpenRefine (please complete the following information):** - Version: Trunk (master) **Additional Context** https://www.slf4j.org/codes.html#multiple_bindings Maybe we want to exclude as runtime dependencies in maven ? and only as compile time ?
main
cleanup the multiple bindings describe the bug api is designed to bind with one and only one underlying logging framework at a time if more than one binding is present on the class path will emit a warning listing the location of those bindings to reproduce steps to reproduce the behavior run refine current results creating new workspace directory home thad local share openrefine class path contains multiple bindings found binding in found binding in see for an explanation actual binding is of type starting openrefine beta expected behavior no multiple bindings warning desktop please complete the following information os ubuntu browser version firefox latest jre or jdk version openjdk openrefine please complete the following information version trunk master additional context maybe we want to exclude as runtime dependencies in maven and only as compile time
1
388,469
11,488,085,758
IssuesEvent
2020-02-11 13:16:56
DigitalCampus/django-oppia
https://api.github.com/repos/DigitalCampus/django-oppia
closed
Add management command for clearing temp directories
enhancement good-first-issue low priority
there can be temp dirs under uploads/ & media/
1.0
Add management command for clearing temp directories - there can be temp dirs under uploads/ & media/
non_main
add management command for clearing temp directories there can be temp dirs under uploads media
0
653,824
21,627,659,589
IssuesEvent
2022-05-05 05:50:41
LuanRT/YouTube.js
https://api.github.com/repos/LuanRT/YouTube.js
closed
YouTube Shorts & Sounds
enhancement good first issue priority: medium
Hey @LuanRT, First up great work on YouTube.js - you're crushing it! With the recent addition of YouTube Shorts, I was wondering if it was possible for you to reverse engineer some of the endpoints YouTube is using explicitly for its new Shorts feature. In particular, there are 2 endpoints that are super interesting to us: 1. When you open the YouTube mobile app and click on the (+) Create icon and select "Create a Short" there is a button to "Add sound" which opens a popup that lets the user search for sounds and add them to their favorites. The endpoint for this as far as I can tell is the following: `https://youtubei.googleapis.com/youtubei/v1/sfv/search` and it uses "youtube-sfv" together with the search query. It would be awesome to be able to search for a sound to get back the ID of that sound. 2. When playing a YouTube Short by going into the main 'Shorts' tab on the mobile app, there is a new sound cover on the bottom right side that directs to a new dedicated sound page if a popular sound was used in a short. Primarily this is being used for big songs that are being played in the shorts. When clicking on the cover artwork and being redirected to that sound page it has a lot of interesting information, like the amount of shorts that were created using that sound, the top shorts videos using the sound, and also a redirect link to the original YouTube video that sound originated from. The endpoint for this should be following: `https://youtubei.googleapis.com/youtubei/v1/browse` and it uses "FEsfv_audio_pivot" together with the ID of the sound. For us all of this information is super valuable as we would like to track these stats and display them to artist and record labels so they can monitor how their tracks are performing on YouTube. Do you think you could help us getting this set up? Happy as well to talk privately and compensate you for your work on this! Please let me know! All the best, Oskar
1.0
YouTube Shorts & Sounds - Hey @LuanRT, First up great work on YouTube.js - you're crushing it! With the recent addition of YouTube Shorts, I was wondering if it was possible for you to reverse engineer some of the endpoints YouTube is using explicitly for its new Shorts feature. In particular, there are 2 endpoints that are super interesting to us: 1. When you open the YouTube mobile app and click on the (+) Create icon and select "Create a Short" there is a button to "Add sound" which opens a popup that lets the user search for sounds and add them to their favorites. The endpoint for this as far as I can tell is the following: `https://youtubei.googleapis.com/youtubei/v1/sfv/search` and it uses "youtube-sfv" together with the search query. It would be awesome to be able to search for a sound to get back the ID of that sound. 2. When playing a YouTube Short by going into the main 'Shorts' tab on the mobile app, there is a new sound cover on the bottom right side that directs to a new dedicated sound page if a popular sound was used in a short. Primarily this is being used for big songs that are being played in the shorts. When clicking on the cover artwork and being redirected to that sound page it has a lot of interesting information, like the amount of shorts that were created using that sound, the top shorts videos using the sound, and also a redirect link to the original YouTube video that sound originated from. The endpoint for this should be following: `https://youtubei.googleapis.com/youtubei/v1/browse` and it uses "FEsfv_audio_pivot" together with the ID of the sound. For us all of this information is super valuable as we would like to track these stats and display them to artist and record labels so they can monitor how their tracks are performing on YouTube. Do you think you could help us getting this set up? Happy as well to talk privately and compensate you for your work on this! Please let me know! All the best, Oskar
non_main
youtube shorts sounds hey luanrt first up great work on youtube js you re crushing it with the recent addition of youtube shorts i was wondering if it was possible for you to reverse engineer some of the endpoints youtube is using explicitly for its new shorts feature in particular there are endpoints that are super interesting to us when you open the youtube mobile app and click on the create icon and select create a short there is a button to add sound which opens a popup that lets the user search for sounds and add them to their favorites the endpoint for this as far as i can tell is the following and it uses youtube sfv together with the search query it would be awesome to be able to search for a sound to get back the id of that sound when playing a youtube short by going into the main shorts tab on the mobile app there is a new sound cover on the bottom right side that directs to a new dedicated sound page if a popular sound was used in a short primarily this is being used for big songs that are being played in the shorts when clicking on the cover artwork and being redirected to that sound page it has a lot of interesting information like the amount of shorts that were created using that sound the top shorts videos using the sound and also a redirect link to the original youtube video that sound originated from the endpoint for this should be following and it uses fesfv audio pivot together with the id of the sound for us all of this information is super valuable as we would like to track these stats and display them to artist and record labels so they can monitor how their tracks are performing on youtube do you think you could help us getting this set up happy as well to talk privately and compensate you for your work on this please let me know all the best oskar
0
47,215
11,984,257,776
IssuesEvent
2020-04-07 15:36:03
Exawind/nalu-wind
https://api.github.com/repos/Exawind/nalu-wind
closed
./abl_mesh -i nalu_abl_mesh.yaml outputs nothing and does not exit with latest wind-utils executables
build-issues
Hello, I'm trying to build nalu-wind with wind-utils to perform simulation with yaw misalignment. I follow the installation manual of Development Build of Nalu-Wind(https://nalu-wind.readthedocs.io/en/latest/source/user/build_spack.html#development-build-of-nalu-wind) cmake -DTrilinos_DIR:PATH=$(spack location -i trilinos) \ -DYAML_DIR:PATH=$(spack location -i yaml-cpp) \ -DCMAKE_BUILD_TYPE=RELEASE \ .. make with -DENABLE_WIND_UTILS=ON added to enable wind utils But met the following error: CMakeFiles/nalu_preprocess.dir/nalu_preprocess.cpp.o:(.data.rel.ro._ZTVN5boost15program_options11typed_valueINSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEcEE[_ZTVN5boost15program_options11typed_valueINSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEcEE]+0x38): undefined reference to `boost::program_options::value_semantic_codecvt_helper<char>::parse(boost::any&, std::vector<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::allocator<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > const&, bool) const' collect2: error: ld returned 1 exit status wind-utils/src/preprocessing/CMakeFiles/nalu_preprocess.dir/build.make:807: recipe for target 'wind-utils/src/preprocessing/nalu_preprocess' failed make[2]: *** [wind-utils/src/preprocessing/nalu_preprocess] Error 1 CMakeFiles/Makefile2:1783: recipe for target 'wind-utils/src/preprocessing/CMakeFiles/nalu_preprocess.dir/all' failed make[1]: *** [wind-utils/src/preprocessing/CMakeFiles/nalu_preprocess.dir/all] Error 2 Makefile:129: recipe for target 'all' failed make: *** [all] Error 2 Is there anyone could help me? Millions of thanks in advance.
1.0
./abl_mesh -i nalu_abl_mesh.yaml outputs nothing and does not exit with latest wind-utils executables - Hello, I'm trying to build nalu-wind with wind-utils to perform simulation with yaw misalignment. I follow the installation manual of Development Build of Nalu-Wind(https://nalu-wind.readthedocs.io/en/latest/source/user/build_spack.html#development-build-of-nalu-wind) cmake -DTrilinos_DIR:PATH=$(spack location -i trilinos) \ -DYAML_DIR:PATH=$(spack location -i yaml-cpp) \ -DCMAKE_BUILD_TYPE=RELEASE \ .. make with -DENABLE_WIND_UTILS=ON added to enable wind utils But met the following error: CMakeFiles/nalu_preprocess.dir/nalu_preprocess.cpp.o:(.data.rel.ro._ZTVN5boost15program_options11typed_valueINSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEcEE[_ZTVN5boost15program_options11typed_valueINSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEcEE]+0x38): undefined reference to `boost::program_options::value_semantic_codecvt_helper<char>::parse(boost::any&, std::vector<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::allocator<std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > > > const&, bool) const' collect2: error: ld returned 1 exit status wind-utils/src/preprocessing/CMakeFiles/nalu_preprocess.dir/build.make:807: recipe for target 'wind-utils/src/preprocessing/nalu_preprocess' failed make[2]: *** [wind-utils/src/preprocessing/nalu_preprocess] Error 1 CMakeFiles/Makefile2:1783: recipe for target 'wind-utils/src/preprocessing/CMakeFiles/nalu_preprocess.dir/all' failed make[1]: *** [wind-utils/src/preprocessing/CMakeFiles/nalu_preprocess.dir/all] Error 2 Makefile:129: recipe for target 'all' failed make: *** [all] Error 2 Is there anyone could help me? Millions of thanks in advance.
non_main
abl mesh i nalu abl mesh yaml outputs nothing and does not exit with latest wind utils executables hello i m trying to build nalu wind with wind utils to perform simulation with yaw misalignment i follow the installation manual of development build of nalu wind cmake dtrilinos dir path spack location i trilinos dyaml dir path spack location i yaml cpp dcmake build type release make with denable wind utils on added to enable wind utils but met the following error cmakefiles nalu preprocess dir nalu preprocess cpp o data rel ro traitsicesaiceeecee undefined reference to boost program options value semantic codecvt helper parse boost any std vector std allocator std allocator std allocator const bool const error ld returned exit status wind utils src preprocessing cmakefiles nalu preprocess dir build make recipe for target wind utils src preprocessing nalu preprocess failed make error cmakefiles recipe for target wind utils src preprocessing cmakefiles nalu preprocess dir all failed make error makefile recipe for target all failed make error is there anyone could help me millions of thanks in advance
0
5,420
27,209,158,855
IssuesEvent
2023-02-20 15:14:48
tom-texier/ToDo-Co
https://api.github.com/repos/tom-texier/ToDo-Co
opened
Audit
maintainability
Produire un audit de code sur les deux axes suivants : **(AVANT ET APRÈS MODIFICATION)** - La qualité du code - La performance _**Qualité :** Codacy ou CodeClimate **Performance :** Profiler de Symfony, Blackfire ou New Relic_
True
Audit - Produire un audit de code sur les deux axes suivants : **(AVANT ET APRÈS MODIFICATION)** - La qualité du code - La performance _**Qualité :** Codacy ou CodeClimate **Performance :** Profiler de Symfony, Blackfire ou New Relic_
main
audit produire un audit de code sur les deux axes suivants avant et après modification la qualité du code la performance qualité codacy ou codeclimate performance profiler de symfony blackfire ou new relic
1
342,593
24,749,977,424
IssuesEvent
2022-10-21 12:59:46
goranagojic/retina-subjective-assessment
https://api.github.com/repos/goranagojic/retina-subjective-assessment
closed
Update README for survey type 1
documentation
Make the following updates: - Create a pipeline to principally show how is survey data transformed into the final result file that is used for statistical analysis. - Describe the format for each of the files from the pipeline.
1.0
Update README for survey type 1 - Make the following updates: - Create a pipeline to principally show how is survey data transformed into the final result file that is used for statistical analysis. - Describe the format for each of the files from the pipeline.
non_main
update readme for survey type make the following updates create a pipeline to principally show how is survey data transformed into the final result file that is used for statistical analysis describe the format for each of the files from the pipeline
0
3,971
18,258,211,701
IssuesEvent
2021-10-03 12:00:14
gama-platform/gama
https://api.github.com/repos/gama-platform/gama
closed
Some update sites do not exist anymore
🛠 Affects Maintainability
![image](https://user-images.githubusercontent.com/6105384/135634046-98944263-876c-4e3b-84d7-13f7bf9d121c.png) It seems above update sites are disconnected. We should remove or resurrect them?
True
Some update sites do not exist anymore - ![image](https://user-images.githubusercontent.com/6105384/135634046-98944263-876c-4e3b-84d7-13f7bf9d121c.png) It seems above update sites are disconnected. We should remove or resurrect them?
main
some update sites do not exist anymore it seems above update sites are disconnected we should remove or resurrect them
1
1,417
6,177,426,441
IssuesEvent
2017-07-02 00:04:08
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Error installing casks that print EULA during dmg mounting: invalid byte sequence in UTF-8
awaiting maintainer feedback bug
**Update.** The problem isn't really a problem of non-ASCII artifacts. Rather, it's the problem of EULA being printed during mounting. See https://github.com/caskroom/homebrew-cask/issues/13402#issuecomment-136195788 for diagnostics. --- I've always had the same issue as #7303, but the issue (being old) was closed last month, so I'm opening a new one. The error occurs to me whenever there are paths containing non-ASCII characters, e.g., with `baiducloud`, `sogouinput`, etc. Curiously, whenever I install the same problematic cask twice in succession, the first attempt will fail with "invalid byte sequence in UTF-8", but the second attempt will succeed. Debug output is given below. Note that the only relevant locale bit in my env is `LC_ALL=en_US.UTF-8`, but it doesn't really matter. I can unset `LC_ALL` and still reproduce the exact same issue. And I can reproduce on vanilla installs of OS X 10.9 (if memory serves, this one has been a while), 10.10, and 10.11. ``` > brew cask install --debug baiducloud ==> Creating directories ==> Loading Cask definitions ==> Translating 'baiducloud' into a valid Cask source ==> Testing source class Hbc::Source::URI ==> Testing source class Hbc::Source::PathSlashRequired ==> Testing source class Hbc::Source::TappedQualified ==> Testing source class Hbc::Source::UntappedQualified ==> Testing source class Hbc::Source::Tapped ==> Success! Using source class Hbc::Source::Tapped ==> Resolved Cask URI or file source to '/usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb' ==> Cask instance dumps in YAML: ==> Cask instance toplevel: --- !ruby/object:KlassPrefixBaiducloud sourcefile_path: !ruby/object:Pathname path: /usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb token: baiducloud ==> Cask instance method 'name': --- - 百度云同步盘 - Baidu Yun Tong Bu Pan - Baigu Cloud ==> Cask instance method 'homepage': --- http://pan.baidu.com ... ==> Cask instance method 'url': --- !ruby/object:Hbc::URL uri: !ruby/object:URI::HTTP scheme: http user: password: host: bcscdn.baidu.com port: 80 path: /netdisk/BaiduYun_2.4.4.dmg query: opaque: registry: fragment: parser: user_agent: cookies: referer: using: revision: trust_cert: data: ==> Cask instance method 'appcast': --- ... ==> Cask instance method 'version': --- 2.4.4 ... ==> Cask instance method 'license': --- !ruby/object:Hbc::DSL::License value: :gratis ==> Cask instance method 'tags': --- ... ==> Cask instance method 'sha256': --- 0a1a438f9ab69d67f0d53d585f0a8c5db7fdbb84ab0c0df55fc06068a67b8149 ... ==> Cask instance method 'artifacts': --- :app: !ruby/object:Set hash: ? - 百度云同步盘.app : true :uninstall: !ruby/object:Set hash: ? :quit: com.baidu.netdiskmac : true :zap: !ruby/object:Set hash: ? :delete: - ~/Library/Application Support/百度云同步盘 - ~/Library/Preferences/com.baidu.netdiskmac.plist - ~/Library/Caches/com.baidu.netdiskmac : true ==> Cask instance method 'caveats': --- [] ==> Cask instance method 'depends_on': --- ... ==> Cask instance method 'conflicts_with': --- ... ==> Cask instance method 'container': --- ... ==> Cask instance method 'gpg': --- ... ==> Cask instance method 'accessibility_access': --- ... ==> Hbc::Installer.install ==> Printing caveats ==> Downloading ==> Downloading http://bcscdn.baidu.com/netdisk/BaiduYun_2.4.4.dmg Already downloaded: /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Downloaded to -> /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Extracting primary container ==> Determining which containers to use based on filetype ==> Checking container class Hbc::Container::Pkg ==> Checking container class Hbc::Container::Ttf ==> Checking container class Hbc::Container::Otf ==> Checking container class Hbc::Container::Air ==> Checking container class Hbc::Container::Cab ==> Executing: ["/usr/bin/file", "-Izb", "--", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Checking container class Hbc::Container::Dmg ==> Executing: ["/usr/bin/hdiutil", "imageinfo", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Using container class Hbc::Container::Dmg for /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> Executing: ["/usr/bin/hdiutil", "mount", "-plist", "-nobrowse", "-readonly", "-noidme", "-mountrandom", "/tmp", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Purging files for version 2.4.4 of Cask baiducloud Error: invalid byte sequence in UTF-8 Most likely, this means you have an outdated version of homebrew-cask. Please run: brew update && brew upgrade brew-cask && brew cleanup && brew cask cleanup If this doesn’t fix the problem, please report this bug: https://github.com/caskroom/homebrew-cask/issues /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/system_command.rb:98:in `_parse_plist' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/system_command.rb:73:in `plist' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/container/dmg.rb:38:in `mount!' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/container/dmg.rb:13:in `extract' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/installer.rb:110:in `extract_primary_container' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/installer.rb:65:in `install' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:20:in `block in install_casks' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:17:in `each' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:17:in `install_casks' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:6:in `run' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli.rb:79:in `run_command' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli.rb:119:in `process' /usr/local/Cellar/brew-cask/0.56.0/rubylib/brew-cask-cmd.rb:18:in `<main>' ``` Second try: ``` > brew cask install --debug baiducloud ==> Creating directories ==> Loading Cask definitions ==> Translating 'baiducloud' into a valid Cask source ==> Testing source class Hbc::Source::URI ==> Testing source class Hbc::Source::PathSlashRequired ==> Testing source class Hbc::Source::TappedQualified ==> Testing source class Hbc::Source::UntappedQualified ==> Testing source class Hbc::Source::Tapped ==> Success! Using source class Hbc::Source::Tapped ==> Resolved Cask URI or file source to '/usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb' ==> Cask instance dumps in YAML: ==> Cask instance toplevel: --- !ruby/object:KlassPrefixBaiducloud sourcefile_path: !ruby/object:Pathname path: /usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb token: baiducloud ==> Cask instance method 'name': --- - 百度云同步盘 - Baidu Yun Tong Bu Pan - Baigu Cloud ==> Cask instance method 'homepage': --- http://pan.baidu.com ... ==> Cask instance method 'url': --- !ruby/object:Hbc::URL uri: !ruby/object:URI::HTTP scheme: http user: password: host: bcscdn.baidu.com port: 80 path: /netdisk/BaiduYun_2.4.4.dmg query: opaque: registry: fragment: parser: user_agent: cookies: referer: using: revision: trust_cert: data: ==> Cask instance method 'appcast': --- ... ==> Cask instance method 'version': --- 2.4.4 ... ==> Cask instance method 'license': --- !ruby/object:Hbc::DSL::License value: :gratis ==> Cask instance method 'tags': --- ... ==> Cask instance method 'sha256': --- 0a1a438f9ab69d67f0d53d585f0a8c5db7fdbb84ab0c0df55fc06068a67b8149 ... ==> Cask instance method 'artifacts': --- :app: !ruby/object:Set hash: ? - 百度云同步盘.app : true :uninstall: !ruby/object:Set hash: ? :quit: com.baidu.netdiskmac : true :zap: !ruby/object:Set hash: ? :delete: - ~/Library/Application Support/百度云同步盘 - ~/Library/Preferences/com.baidu.netdiskmac.plist - ~/Library/Caches/com.baidu.netdiskmac : true ==> Cask instance method 'caveats': --- [] ==> Cask instance method 'depends_on': --- ... ==> Cask instance method 'conflicts_with': --- ... ==> Cask instance method 'container': --- ... ==> Cask instance method 'gpg': --- ... ==> Cask instance method 'accessibility_access': --- ... ==> Hbc::Installer.install ==> Printing caveats ==> Downloading ==> Downloading http://bcscdn.baidu.com/netdisk/BaiduYun_2.4.4.dmg Already downloaded: /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Downloaded to -> /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Extracting primary container ==> Determining which containers to use based on filetype ==> Checking container class Hbc::Container::Pkg ==> Checking container class Hbc::Container::Ttf ==> Checking container class Hbc::Container::Otf ==> Checking container class Hbc::Container::Air ==> Checking container class Hbc::Container::Cab ==> Executing: ["/usr/bin/file", "-Izb", "--", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Checking container class Hbc::Container::Dmg ==> Executing: ["/usr/bin/hdiutil", "imageinfo", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Using container class Hbc::Container::Dmg for /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> Executing: ["/usr/bin/hdiutil", "mount", "-plist", "-nobrowse", "-readonly", "-noidme", "-mountrandom", "/tmp", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Executing: ["/usr/bin/ditto", "--", "/private/tmp/dmg.8goCsO", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4>"] ==> Executing: ["/usr/sbin/diskutil", "eject", "#<Pathname:/private/tmp/dmg.8goCsO>"] ==> Installing artifacts ==> Determining which artifacts are present in Cask baiducloud ==> Checking for artifact class Hbc::Artifact::PreflightBlock ==> Checking for artifact class Hbc::Artifact::NestedContainer ==> Checking for artifact class Hbc::Artifact::Installer ==> Checking for artifact class Hbc::Artifact::App ==> Checking for artifact class Hbc::Artifact::Suite ==> Checking for artifact class Hbc::Artifact::Artifact ==> Checking for artifact class Hbc::Artifact::Colorpicker ==> Checking for artifact class Hbc::Artifact::Pkg ==> Checking for artifact class Hbc::Artifact::Prefpane ==> Checking for artifact class Hbc::Artifact::Qlplugin ==> Checking for artifact class Hbc::Artifact::Font ==> Checking for artifact class Hbc::Artifact::Service ==> Checking for artifact class Hbc::Artifact::StageOnly ==> Checking for artifact class Hbc::Artifact::Binary ==> Checking for artifact class Hbc::Artifact::InputMethod ==> Checking for artifact class Hbc::Artifact::InternetPlugin ==> Checking for artifact class Hbc::Artifact::ScreenSaver ==> Checking for artifact class Hbc::Artifact::Uninstall ==> Checking for artifact class Hbc::Artifact::PostflightBlock ==> Checking for artifact class Hbc::Artifact::Zap ==> 3 artifact/s defined Hbc::Artifact::App Hbc::Artifact::Uninstall Hbc::Artifact::Zap ==> Installing artifact of class Hbc::Artifact::App ==> Symlinking App '百度云同步盘.app' to '/Applications/百度云同步盘.app' ==> Executing: ["/bin/ln", "-hfs", "--", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4/百度云同步盘.app>", "#<Pathname:/Applications/百度云同步盘.app>"] ==> Installing artifact of class Hbc::Artifact::Uninstall ==> Nothing to do. The uninstall artifact has no install phase. ==> Installing artifact of class Hbc::Artifact::Zap ==> Nothing to do. The zap artifact has no install phase. ==> Creating metadata directory /opt/homebrew-cask/Caskroom/baiducloud/.metadata/2.4.4/20150824053746.549 ==> Creating metadata subdirectory /opt/homebrew-cask/Caskroom/baiducloud/.metadata/2.4.4/20150824053746.549/Casks ==> Executing: ["/usr/bin/find", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4>", "-type", "f", "-not", "-name", ".DS_Store", "-print0"] ==> Executing: ["/usr/bin/du", "-hs", "--", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4>"] 🍺 baiducloud staged at '/opt/homebrew-cask/Caskroom/baiducloud/2.4.4' (118 files, 7.5M) ``` --- Update: `brew doctor` and `brew cask doctor` output: https://gist.github.com/zmwangx/c5605892a4f990bf37f9.
True
Error installing casks that print EULA during dmg mounting: invalid byte sequence in UTF-8 - **Update.** The problem isn't really a problem of non-ASCII artifacts. Rather, it's the problem of EULA being printed during mounting. See https://github.com/caskroom/homebrew-cask/issues/13402#issuecomment-136195788 for diagnostics. --- I've always had the same issue as #7303, but the issue (being old) was closed last month, so I'm opening a new one. The error occurs to me whenever there are paths containing non-ASCII characters, e.g., with `baiducloud`, `sogouinput`, etc. Curiously, whenever I install the same problematic cask twice in succession, the first attempt will fail with "invalid byte sequence in UTF-8", but the second attempt will succeed. Debug output is given below. Note that the only relevant locale bit in my env is `LC_ALL=en_US.UTF-8`, but it doesn't really matter. I can unset `LC_ALL` and still reproduce the exact same issue. And I can reproduce on vanilla installs of OS X 10.9 (if memory serves, this one has been a while), 10.10, and 10.11. ``` > brew cask install --debug baiducloud ==> Creating directories ==> Loading Cask definitions ==> Translating 'baiducloud' into a valid Cask source ==> Testing source class Hbc::Source::URI ==> Testing source class Hbc::Source::PathSlashRequired ==> Testing source class Hbc::Source::TappedQualified ==> Testing source class Hbc::Source::UntappedQualified ==> Testing source class Hbc::Source::Tapped ==> Success! Using source class Hbc::Source::Tapped ==> Resolved Cask URI or file source to '/usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb' ==> Cask instance dumps in YAML: ==> Cask instance toplevel: --- !ruby/object:KlassPrefixBaiducloud sourcefile_path: !ruby/object:Pathname path: /usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb token: baiducloud ==> Cask instance method 'name': --- - 百度云同步盘 - Baidu Yun Tong Bu Pan - Baigu Cloud ==> Cask instance method 'homepage': --- http://pan.baidu.com ... ==> Cask instance method 'url': --- !ruby/object:Hbc::URL uri: !ruby/object:URI::HTTP scheme: http user: password: host: bcscdn.baidu.com port: 80 path: /netdisk/BaiduYun_2.4.4.dmg query: opaque: registry: fragment: parser: user_agent: cookies: referer: using: revision: trust_cert: data: ==> Cask instance method 'appcast': --- ... ==> Cask instance method 'version': --- 2.4.4 ... ==> Cask instance method 'license': --- !ruby/object:Hbc::DSL::License value: :gratis ==> Cask instance method 'tags': --- ... ==> Cask instance method 'sha256': --- 0a1a438f9ab69d67f0d53d585f0a8c5db7fdbb84ab0c0df55fc06068a67b8149 ... ==> Cask instance method 'artifacts': --- :app: !ruby/object:Set hash: ? - 百度云同步盘.app : true :uninstall: !ruby/object:Set hash: ? :quit: com.baidu.netdiskmac : true :zap: !ruby/object:Set hash: ? :delete: - ~/Library/Application Support/百度云同步盘 - ~/Library/Preferences/com.baidu.netdiskmac.plist - ~/Library/Caches/com.baidu.netdiskmac : true ==> Cask instance method 'caveats': --- [] ==> Cask instance method 'depends_on': --- ... ==> Cask instance method 'conflicts_with': --- ... ==> Cask instance method 'container': --- ... ==> Cask instance method 'gpg': --- ... ==> Cask instance method 'accessibility_access': --- ... ==> Hbc::Installer.install ==> Printing caveats ==> Downloading ==> Downloading http://bcscdn.baidu.com/netdisk/BaiduYun_2.4.4.dmg Already downloaded: /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Downloaded to -> /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Extracting primary container ==> Determining which containers to use based on filetype ==> Checking container class Hbc::Container::Pkg ==> Checking container class Hbc::Container::Ttf ==> Checking container class Hbc::Container::Otf ==> Checking container class Hbc::Container::Air ==> Checking container class Hbc::Container::Cab ==> Executing: ["/usr/bin/file", "-Izb", "--", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Checking container class Hbc::Container::Dmg ==> Executing: ["/usr/bin/hdiutil", "imageinfo", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Using container class Hbc::Container::Dmg for /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> Executing: ["/usr/bin/hdiutil", "mount", "-plist", "-nobrowse", "-readonly", "-noidme", "-mountrandom", "/tmp", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Purging files for version 2.4.4 of Cask baiducloud Error: invalid byte sequence in UTF-8 Most likely, this means you have an outdated version of homebrew-cask. Please run: brew update && brew upgrade brew-cask && brew cleanup && brew cask cleanup If this doesn’t fix the problem, please report this bug: https://github.com/caskroom/homebrew-cask/issues /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/system_command.rb:98:in `_parse_plist' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/system_command.rb:73:in `plist' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/container/dmg.rb:38:in `mount!' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/container/dmg.rb:13:in `extract' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/installer.rb:110:in `extract_primary_container' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/installer.rb:65:in `install' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:20:in `block in install_casks' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:17:in `each' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:17:in `install_casks' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli/install.rb:6:in `run' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli.rb:79:in `run_command' /usr/local/Cellar/brew-cask/0.56.0/rubylib/hbc/cli.rb:119:in `process' /usr/local/Cellar/brew-cask/0.56.0/rubylib/brew-cask-cmd.rb:18:in `<main>' ``` Second try: ``` > brew cask install --debug baiducloud ==> Creating directories ==> Loading Cask definitions ==> Translating 'baiducloud' into a valid Cask source ==> Testing source class Hbc::Source::URI ==> Testing source class Hbc::Source::PathSlashRequired ==> Testing source class Hbc::Source::TappedQualified ==> Testing source class Hbc::Source::UntappedQualified ==> Testing source class Hbc::Source::Tapped ==> Success! Using source class Hbc::Source::Tapped ==> Resolved Cask URI or file source to '/usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb' ==> Cask instance dumps in YAML: ==> Cask instance toplevel: --- !ruby/object:KlassPrefixBaiducloud sourcefile_path: !ruby/object:Pathname path: /usr/local/Library/Taps/caskroom/homebrew-cask/Casks/baiducloud.rb token: baiducloud ==> Cask instance method 'name': --- - 百度云同步盘 - Baidu Yun Tong Bu Pan - Baigu Cloud ==> Cask instance method 'homepage': --- http://pan.baidu.com ... ==> Cask instance method 'url': --- !ruby/object:Hbc::URL uri: !ruby/object:URI::HTTP scheme: http user: password: host: bcscdn.baidu.com port: 80 path: /netdisk/BaiduYun_2.4.4.dmg query: opaque: registry: fragment: parser: user_agent: cookies: referer: using: revision: trust_cert: data: ==> Cask instance method 'appcast': --- ... ==> Cask instance method 'version': --- 2.4.4 ... ==> Cask instance method 'license': --- !ruby/object:Hbc::DSL::License value: :gratis ==> Cask instance method 'tags': --- ... ==> Cask instance method 'sha256': --- 0a1a438f9ab69d67f0d53d585f0a8c5db7fdbb84ab0c0df55fc06068a67b8149 ... ==> Cask instance method 'artifacts': --- :app: !ruby/object:Set hash: ? - 百度云同步盘.app : true :uninstall: !ruby/object:Set hash: ? :quit: com.baidu.netdiskmac : true :zap: !ruby/object:Set hash: ? :delete: - ~/Library/Application Support/百度云同步盘 - ~/Library/Preferences/com.baidu.netdiskmac.plist - ~/Library/Caches/com.baidu.netdiskmac : true ==> Cask instance method 'caveats': --- [] ==> Cask instance method 'depends_on': --- ... ==> Cask instance method 'conflicts_with': --- ... ==> Cask instance method 'container': --- ... ==> Cask instance method 'gpg': --- ... ==> Cask instance method 'accessibility_access': --- ... ==> Hbc::Installer.install ==> Printing caveats ==> Downloading ==> Downloading http://bcscdn.baidu.com/netdisk/BaiduYun_2.4.4.dmg Already downloaded: /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Downloaded to -> /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> SHA256 checksums match ==> Extracting primary container ==> Determining which containers to use based on filetype ==> Checking container class Hbc::Container::Pkg ==> Checking container class Hbc::Container::Ttf ==> Checking container class Hbc::Container::Otf ==> Checking container class Hbc::Container::Air ==> Checking container class Hbc::Container::Cab ==> Executing: ["/usr/bin/file", "-Izb", "--", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Checking container class Hbc::Container::Dmg ==> Executing: ["/usr/bin/hdiutil", "imageinfo", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Using container class Hbc::Container::Dmg for /Library/Caches/Homebrew/baiducloud-2.4.4.dmg ==> Executing: ["/usr/bin/hdiutil", "mount", "-plist", "-nobrowse", "-readonly", "-noidme", "-mountrandom", "/tmp", "#<Pathname:/Library/Caches/Homebrew/baiducloud-2.4.4.dmg>"] ==> Executing: ["/usr/bin/ditto", "--", "/private/tmp/dmg.8goCsO", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4>"] ==> Executing: ["/usr/sbin/diskutil", "eject", "#<Pathname:/private/tmp/dmg.8goCsO>"] ==> Installing artifacts ==> Determining which artifacts are present in Cask baiducloud ==> Checking for artifact class Hbc::Artifact::PreflightBlock ==> Checking for artifact class Hbc::Artifact::NestedContainer ==> Checking for artifact class Hbc::Artifact::Installer ==> Checking for artifact class Hbc::Artifact::App ==> Checking for artifact class Hbc::Artifact::Suite ==> Checking for artifact class Hbc::Artifact::Artifact ==> Checking for artifact class Hbc::Artifact::Colorpicker ==> Checking for artifact class Hbc::Artifact::Pkg ==> Checking for artifact class Hbc::Artifact::Prefpane ==> Checking for artifact class Hbc::Artifact::Qlplugin ==> Checking for artifact class Hbc::Artifact::Font ==> Checking for artifact class Hbc::Artifact::Service ==> Checking for artifact class Hbc::Artifact::StageOnly ==> Checking for artifact class Hbc::Artifact::Binary ==> Checking for artifact class Hbc::Artifact::InputMethod ==> Checking for artifact class Hbc::Artifact::InternetPlugin ==> Checking for artifact class Hbc::Artifact::ScreenSaver ==> Checking for artifact class Hbc::Artifact::Uninstall ==> Checking for artifact class Hbc::Artifact::PostflightBlock ==> Checking for artifact class Hbc::Artifact::Zap ==> 3 artifact/s defined Hbc::Artifact::App Hbc::Artifact::Uninstall Hbc::Artifact::Zap ==> Installing artifact of class Hbc::Artifact::App ==> Symlinking App '百度云同步盘.app' to '/Applications/百度云同步盘.app' ==> Executing: ["/bin/ln", "-hfs", "--", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4/百度云同步盘.app>", "#<Pathname:/Applications/百度云同步盘.app>"] ==> Installing artifact of class Hbc::Artifact::Uninstall ==> Nothing to do. The uninstall artifact has no install phase. ==> Installing artifact of class Hbc::Artifact::Zap ==> Nothing to do. The zap artifact has no install phase. ==> Creating metadata directory /opt/homebrew-cask/Caskroom/baiducloud/.metadata/2.4.4/20150824053746.549 ==> Creating metadata subdirectory /opt/homebrew-cask/Caskroom/baiducloud/.metadata/2.4.4/20150824053746.549/Casks ==> Executing: ["/usr/bin/find", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4>", "-type", "f", "-not", "-name", ".DS_Store", "-print0"] ==> Executing: ["/usr/bin/du", "-hs", "--", "#<Pathname:/opt/homebrew-cask/Caskroom/baiducloud/2.4.4>"] 🍺 baiducloud staged at '/opt/homebrew-cask/Caskroom/baiducloud/2.4.4' (118 files, 7.5M) ``` --- Update: `brew doctor` and `brew cask doctor` output: https://gist.github.com/zmwangx/c5605892a4f990bf37f9.
main
error installing casks that print eula during dmg mounting invalid byte sequence in utf update the problem isn t really a problem of non ascii artifacts rather it s the problem of eula being printed during mounting see for diagnostics i ve always had the same issue as but the issue being old was closed last month so i m opening a new one the error occurs to me whenever there are paths containing non ascii characters e g with baiducloud sogouinput etc curiously whenever i install the same problematic cask twice in succession the first attempt will fail with invalid byte sequence in utf but the second attempt will succeed debug output is given below note that the only relevant locale bit in my env is lc all en us utf but it doesn t really matter i can unset lc all and still reproduce the exact same issue and i can reproduce on vanilla installs of os x if memory serves this one has been a while and brew cask install debug baiducloud creating directories loading cask definitions translating baiducloud into a valid cask source testing source class hbc source uri testing source class hbc source pathslashrequired testing source class hbc source tappedqualified testing source class hbc source untappedqualified testing source class hbc source tapped success using source class hbc source tapped resolved cask uri or file source to usr local library taps caskroom homebrew cask casks baiducloud rb cask instance dumps in yaml cask instance toplevel ruby object klassprefixbaiducloud sourcefile path ruby object pathname path usr local library taps caskroom homebrew cask casks baiducloud rb token baiducloud cask instance method name 百度云同步盘 baidu yun tong bu pan baigu cloud cask instance method homepage cask instance method url ruby object hbc url uri ruby object uri http scheme http user password host bcscdn baidu com port path netdisk baiduyun dmg query opaque registry fragment parser user agent cookies referer using revision trust cert data cask instance method appcast cask instance method version cask instance method license ruby object hbc dsl license value gratis cask instance method tags cask instance method cask instance method artifacts app ruby object set hash 百度云同步盘 app true uninstall ruby object set hash quit com baidu netdiskmac true zap ruby object set hash delete library application support 百度云同步盘 library preferences com baidu netdiskmac plist library caches com baidu netdiskmac true cask instance method caveats cask instance method depends on cask instance method conflicts with cask instance method container cask instance method gpg cask instance method accessibility access hbc installer install printing caveats downloading downloading already downloaded library caches homebrew baiducloud dmg checksums match downloaded to library caches homebrew baiducloud dmg checksums match extracting primary container determining which containers to use based on filetype checking container class hbc container pkg checking container class hbc container ttf checking container class hbc container otf checking container class hbc container air checking container class hbc container cab executing checking container class hbc container dmg executing using container class hbc container dmg for library caches homebrew baiducloud dmg executing purging files for version of cask baiducloud error invalid byte sequence in utf most likely this means you have an outdated version of homebrew cask please run brew update brew upgrade brew cask brew cleanup brew cask cleanup if this doesn’t fix the problem please report this bug usr local cellar brew cask rubylib hbc system command rb in parse plist usr local cellar brew cask rubylib hbc system command rb in plist usr local cellar brew cask rubylib hbc container dmg rb in mount usr local cellar brew cask rubylib hbc container dmg rb in extract usr local cellar brew cask rubylib hbc installer rb in extract primary container usr local cellar brew cask rubylib hbc installer rb in install usr local cellar brew cask rubylib hbc cli install rb in block in install casks usr local cellar brew cask rubylib hbc cli install rb in each usr local cellar brew cask rubylib hbc cli install rb in install casks usr local cellar brew cask rubylib hbc cli install rb in run usr local cellar brew cask rubylib hbc cli rb in run command usr local cellar brew cask rubylib hbc cli rb in process usr local cellar brew cask rubylib brew cask cmd rb in second try brew cask install debug baiducloud creating directories loading cask definitions translating baiducloud into a valid cask source testing source class hbc source uri testing source class hbc source pathslashrequired testing source class hbc source tappedqualified testing source class hbc source untappedqualified testing source class hbc source tapped success using source class hbc source tapped resolved cask uri or file source to usr local library taps caskroom homebrew cask casks baiducloud rb cask instance dumps in yaml cask instance toplevel ruby object klassprefixbaiducloud sourcefile path ruby object pathname path usr local library taps caskroom homebrew cask casks baiducloud rb token baiducloud cask instance method name 百度云同步盘 baidu yun tong bu pan baigu cloud cask instance method homepage cask instance method url ruby object hbc url uri ruby object uri http scheme http user password host bcscdn baidu com port path netdisk baiduyun dmg query opaque registry fragment parser user agent cookies referer using revision trust cert data cask instance method appcast cask instance method version cask instance method license ruby object hbc dsl license value gratis cask instance method tags cask instance method cask instance method artifacts app ruby object set hash 百度云同步盘 app true uninstall ruby object set hash quit com baidu netdiskmac true zap ruby object set hash delete library application support 百度云同步盘 library preferences com baidu netdiskmac plist library caches com baidu netdiskmac true cask instance method caveats cask instance method depends on cask instance method conflicts with cask instance method container cask instance method gpg cask instance method accessibility access hbc installer install printing caveats downloading downloading already downloaded library caches homebrew baiducloud dmg checksums match downloaded to library caches homebrew baiducloud dmg checksums match extracting primary container determining which containers to use based on filetype checking container class hbc container pkg checking container class hbc container ttf checking container class hbc container otf checking container class hbc container air checking container class hbc container cab executing checking container class hbc container dmg executing using container class hbc container dmg for library caches homebrew baiducloud dmg executing executing executing installing artifacts determining which artifacts are present in cask baiducloud checking for artifact class hbc artifact preflightblock checking for artifact class hbc artifact nestedcontainer checking for artifact class hbc artifact installer checking for artifact class hbc artifact app checking for artifact class hbc artifact suite checking for artifact class hbc artifact artifact checking for artifact class hbc artifact colorpicker checking for artifact class hbc artifact pkg checking for artifact class hbc artifact prefpane checking for artifact class hbc artifact qlplugin checking for artifact class hbc artifact font checking for artifact class hbc artifact service checking for artifact class hbc artifact stageonly checking for artifact class hbc artifact binary checking for artifact class hbc artifact inputmethod checking for artifact class hbc artifact internetplugin checking for artifact class hbc artifact screensaver checking for artifact class hbc artifact uninstall checking for artifact class hbc artifact postflightblock checking for artifact class hbc artifact zap artifact s defined hbc artifact app hbc artifact uninstall hbc artifact zap installing artifact of class hbc artifact app symlinking app 百度云同步盘 app to applications 百度云同步盘 app executing installing artifact of class hbc artifact uninstall nothing to do the uninstall artifact has no install phase installing artifact of class hbc artifact zap nothing to do the zap artifact has no install phase creating metadata directory opt homebrew cask caskroom baiducloud metadata creating metadata subdirectory opt homebrew cask caskroom baiducloud metadata casks executing executing 🍺 baiducloud staged at opt homebrew cask caskroom baiducloud files update brew doctor and brew cask doctor output
1
219,006
7,333,047,523
IssuesEvent
2018-03-05 18:08:24
NCEAS/metacat
https://api.github.com/repos/NCEAS/metacat
closed
sql injection bug in EventLog.getD1Record()
Category: metacat Component: Bugzilla-Id Priority: Immediate Status: Rejected Tracker: Bug
--- Author Name: **Matt Jones** (Matt Jones) Original Redmine Issue: 6644, https://projects.ecoinformatics.org/ecoinfo/issues/6644 Original Date: 2015-01-08 Original Assignee: ben leinfelder --- There is a SQL injection bug in getD1Record due to the use of unsantized user input in the database sql command that is constructed via string concatenation. In general, it is never safe to use string concatenation with user input, and as a local function never can really know when function arguments might have originated from an unsafe source, we should always instead use PreparedStatements with parameter bindings. In a previous relelase of Metacat we had expunged all such usages, but it appears this new function was built later. It needs to be refactored ASAP to use PreparedStatements and parameter binding rather than string appends.
1.0
sql injection bug in EventLog.getD1Record() - --- Author Name: **Matt Jones** (Matt Jones) Original Redmine Issue: 6644, https://projects.ecoinformatics.org/ecoinfo/issues/6644 Original Date: 2015-01-08 Original Assignee: ben leinfelder --- There is a SQL injection bug in getD1Record due to the use of unsantized user input in the database sql command that is constructed via string concatenation. In general, it is never safe to use string concatenation with user input, and as a local function never can really know when function arguments might have originated from an unsafe source, we should always instead use PreparedStatements with parameter bindings. In a previous relelase of Metacat we had expunged all such usages, but it appears this new function was built later. It needs to be refactored ASAP to use PreparedStatements and parameter binding rather than string appends.
non_main
sql injection bug in eventlog author name matt jones matt jones original redmine issue original date original assignee ben leinfelder there is a sql injection bug in due to the use of unsantized user input in the database sql command that is constructed via string concatenation in general it is never safe to use string concatenation with user input and as a local function never can really know when function arguments might have originated from an unsafe source we should always instead use preparedstatements with parameter bindings in a previous relelase of metacat we had expunged all such usages but it appears this new function was built later it needs to be refactored asap to use preparedstatements and parameter binding rather than string appends
0
446,197
31,395,507,917
IssuesEvent
2023-08-26 22:08:19
Noravsky/Misiotronica
https://api.github.com/repos/Noravsky/Misiotronica
closed
Obs grales
documentation
Se nota una mejora entre la primer version y la ultima en la rama Dev. El proyecto esta bien encaminado. Es suficiente para avanzar pero necesita trabajar mas en el estilado CSS. Para poder controlar mejor la manera en que presenta el contenido.
1.0
Obs grales - Se nota una mejora entre la primer version y la ultima en la rama Dev. El proyecto esta bien encaminado. Es suficiente para avanzar pero necesita trabajar mas en el estilado CSS. Para poder controlar mejor la manera en que presenta el contenido.
non_main
obs grales se nota una mejora entre la primer version y la ultima en la rama dev el proyecto esta bien encaminado es suficiente para avanzar pero necesita trabajar mas en el estilado css para poder controlar mejor la manera en que presenta el contenido
0
5,329
2,928,287,038
IssuesEvent
2015-06-27 01:58:59
flynn/flynn
https://api.github.com/repos/flynn/flynn
opened
Add a JavaScript style guide
documentation
The basic guidelines are - Run [jshint](http://jshint.com/) - Use tabs instead of spaces - Define vars individually (better indentation with tabs and cleaner diff when making changes) - Use descriptive var names (avoid single character variable names except in for loops) - Use `null` instead of `undefined` - Avoid using `!someVar` except with booleans - Give single space on either side of `!expression` in an `if` statement
1.0
Add a JavaScript style guide - The basic guidelines are - Run [jshint](http://jshint.com/) - Use tabs instead of spaces - Define vars individually (better indentation with tabs and cleaner diff when making changes) - Use descriptive var names (avoid single character variable names except in for loops) - Use `null` instead of `undefined` - Avoid using `!someVar` except with booleans - Give single space on either side of `!expression` in an `if` statement
non_main
add a javascript style guide the basic guidelines are run use tabs instead of spaces define vars individually better indentation with tabs and cleaner diff when making changes use descriptive var names avoid single character variable names except in for loops use null instead of undefined avoid using somevar except with booleans give single space on either side of expression in an if statement
0
44,136
2,899,437,716
IssuesEvent
2015-06-17 11:20:41
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Enhancment for Vesuvio
High Priority Indirect Inelastic SSC
This issue was originally [TRAC 11235](http://trac.mantidproject.org/mantid/ticket/11235) This ticket is blocked by : - [X] [TRAC11205](http://trac.mantidproject.org/mantid/ticket/11205) For startup in June. Martyn please add the blocking subtickets - - - - Keywords: SSC,2015, RAL, indirectShort
1.0
Enhancment for Vesuvio - This issue was originally [TRAC 11235](http://trac.mantidproject.org/mantid/ticket/11235) This ticket is blocked by : - [X] [TRAC11205](http://trac.mantidproject.org/mantid/ticket/11205) For startup in June. Martyn please add the blocking subtickets - - - - Keywords: SSC,2015, RAL, indirectShort
non_main
enhancment for vesuvio this issue was originally this ticket is blocked by for startup in june martyn please add the blocking subtickets keywords ssc ral indirectshort
0
4,929
25,338,604,131
IssuesEvent
2022-11-18 19:11:02
aws/aws-sam-cli-pipeline-init-templates
https://api.github.com/repos/aws/aws-sam-cli-pipeline-init-templates
closed
sam pipeline init for github actions fails on 1st question
bug maintainer/need-response
sam pipeline init erroring out on > Error: 'Invalid question key "shared_values" referenced in default answer of question permissions_provider' Checking for **shared_values** in the repo, indeed it's only ever referenced and isn't defined anywhere
True
sam pipeline init for github actions fails on 1st question - sam pipeline init erroring out on > Error: 'Invalid question key "shared_values" referenced in default answer of question permissions_provider' Checking for **shared_values** in the repo, indeed it's only ever referenced and isn't defined anywhere
main
sam pipeline init for github actions fails on question sam pipeline init erroring out on error invalid question key shared values referenced in default answer of question permissions provider checking for shared values in the repo indeed it s only ever referenced and isn t defined anywhere
1
56,395
15,054,653,988
IssuesEvent
2021-02-03 17:42:51
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
opened
FileUpload: Wrong invalid file name pattern
defect
**Describe the defect** Since 8.0, file name is validated on server side (for security reasons, mostly intrusions). Though, a filename like `D2021-00101/[D-00001742]xxx.jpg` throws an exception while there is no attempt to path traversal. Exception is throw just because there is a `/` **Reproducer** Tricky as I'm uploading a directory using https://github.com/primefaces/primefaces/issues/2705 tweak **Environment:** - PF Version: 8.0 **Expected behavior** No exception. On a general note, I don't think PF should define these validation rule, it's a very tricky topic... Library like this one https://github.com/ESAPI/esapi-java-legacy exists already and seems to do a good job. Also, users should have the possibility to skip those tests, it's company responsability to use whether or not FileUpload safety features. In mine, we don't want it for several reasons (perf, owasp compliancy etc.) Though, I'm thinking of having an attribute `FileUpload#secure` that will validate content type, filename, content etc. _(see Esapi's `DefaultValidator#asserValidFileUpload`)_ Maybe virusScan also?
1.0
FileUpload: Wrong invalid file name pattern - **Describe the defect** Since 8.0, file name is validated on server side (for security reasons, mostly intrusions). Though, a filename like `D2021-00101/[D-00001742]xxx.jpg` throws an exception while there is no attempt to path traversal. Exception is throw just because there is a `/` **Reproducer** Tricky as I'm uploading a directory using https://github.com/primefaces/primefaces/issues/2705 tweak **Environment:** - PF Version: 8.0 **Expected behavior** No exception. On a general note, I don't think PF should define these validation rule, it's a very tricky topic... Library like this one https://github.com/ESAPI/esapi-java-legacy exists already and seems to do a good job. Also, users should have the possibility to skip those tests, it's company responsability to use whether or not FileUpload safety features. In mine, we don't want it for several reasons (perf, owasp compliancy etc.) Though, I'm thinking of having an attribute `FileUpload#secure` that will validate content type, filename, content etc. _(see Esapi's `DefaultValidator#asserValidFileUpload`)_ Maybe virusScan also?
non_main
fileupload wrong invalid file name pattern describe the defect since file name is validated on server side for security reasons mostly intrusions though a filename like xxx jpg throws an exception while there is no attempt to path traversal exception is throw just because there is a reproducer tricky as i m uploading a directory using tweak environment pf version expected behavior no exception on a general note i don t think pf should define these validation rule it s a very tricky topic library like this one exists already and seems to do a good job also users should have the possibility to skip those tests it s company responsability to use whether or not fileupload safety features in mine we don t want it for several reasons perf owasp compliancy etc though i m thinking of having an attribute fileupload secure that will validate content type filename content etc see esapi s defaultvalidator asservalidfileupload maybe virusscan also
0
2,290
8,153,886,783
IssuesEvent
2018-08-23 00:05:19
TravisSpark/spark-website
https://api.github.com/repos/TravisSpark/spark-website
closed
Review and Modify Legacy code-mil files
maintainence
### Checklist - [X] Searched for, and did not find, duplicate [issue](https://github.com/TravisSpark/spark-website/issues) - [X] Indicated whether the issue is a bug or a feature - [X] Focused on one specific bug/feature - [X] Gave a concise and relevant name - [X] Created clear and concise description - [X] Outlined which components are affected - [X] Assigned issue to project, appropriate contributors, and relevant labels <!-- Edit as Appropriate --> ### Issue Type: Feature <!-- Pick one --> ### Description There are legacy files from code.mil. They need to be removed or modified for travisspark.org. 1. robots.txt is set for beta-code-mil (Related to #50 ) 2. serve no longer has a function 3. code.json no longer has a function 4. travis.yml contains commented out code that can be permanently removed 5. screenshot.js no longer has a function ### Affected Components Backend Files in home directory
True
Review and Modify Legacy code-mil files - ### Checklist - [X] Searched for, and did not find, duplicate [issue](https://github.com/TravisSpark/spark-website/issues) - [X] Indicated whether the issue is a bug or a feature - [X] Focused on one specific bug/feature - [X] Gave a concise and relevant name - [X] Created clear and concise description - [X] Outlined which components are affected - [X] Assigned issue to project, appropriate contributors, and relevant labels <!-- Edit as Appropriate --> ### Issue Type: Feature <!-- Pick one --> ### Description There are legacy files from code.mil. They need to be removed or modified for travisspark.org. 1. robots.txt is set for beta-code-mil (Related to #50 ) 2. serve no longer has a function 3. code.json no longer has a function 4. travis.yml contains commented out code that can be permanently removed 5. screenshot.js no longer has a function ### Affected Components Backend Files in home directory
main
review and modify legacy code mil files checklist searched for and did not find duplicate indicated whether the issue is a bug or a feature focused on one specific bug feature gave a concise and relevant name created clear and concise description outlined which components are affected assigned issue to project appropriate contributors and relevant labels issue type feature description there are legacy files from code mil they need to be removed or modified for travisspark org robots txt is set for beta code mil related to serve no longer has a function code json no longer has a function travis yml contains commented out code that can be permanently removed screenshot js no longer has a function affected components backend files in home directory
1
1,749
6,574,943,380
IssuesEvent
2017-09-11 14:34:10
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
shell command not escaping double quotes correctly
affects_2.2 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME shell-module ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 (detached HEAD 44faad0593) last updated 2016/10/18 10:21:47 (GMT +000) ``` ##### OS / ENVIRONMENT Debian 8.6 Linux machine0 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux ##### SUMMARY For some formatting reasons, double curly braces are needed in a shell command. This worked in the past. But how, a hashtag is somehow produces by that workaround. See the steps to reproduce. ##### STEPS TO REPRODUCE ``` - name: Get version of current docker-engine shell: "/usr/bin/docker version --format '{{ '{{' }}.Client.Version{{ '}}' }}' 2>/dev/null | true" register: installed_docker_version - debug: var=installed_docker_version when: installed_docker_version is defined ``` ##### EXPECTED RESULTS The actual version of docker ##### ACTUAL RESULTS The result is: ``` ok: [machine_1] => { "installed_docker_version": { "changed": false, "cmd": "/usr/bin/docker version --format '{#.Client.Version#}' 2>/dev/null | true", "delta": "0:00:00.012714", "end": "2016-10-18 15:40:21.770477", "rc": 0, "start": "2016-10-18 15:40:21.757763", "stderr": "", "stdout": "", "stdout_lines": [], "warnings": [] } } ``` see the part `cmd` of that command. This is clearly wrong. It is also not just a debugging issue, because the correct command would have a correct result. This one just returns nothing.
True
shell command not escaping double quotes correctly - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME shell-module ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 (detached HEAD 44faad0593) last updated 2016/10/18 10:21:47 (GMT +000) ``` ##### OS / ENVIRONMENT Debian 8.6 Linux machine0 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux ##### SUMMARY For some formatting reasons, double curly braces are needed in a shell command. This worked in the past. But how, a hashtag is somehow produces by that workaround. See the steps to reproduce. ##### STEPS TO REPRODUCE ``` - name: Get version of current docker-engine shell: "/usr/bin/docker version --format '{{ '{{' }}.Client.Version{{ '}}' }}' 2>/dev/null | true" register: installed_docker_version - debug: var=installed_docker_version when: installed_docker_version is defined ``` ##### EXPECTED RESULTS The actual version of docker ##### ACTUAL RESULTS The result is: ``` ok: [machine_1] => { "installed_docker_version": { "changed": false, "cmd": "/usr/bin/docker version --format '{#.Client.Version#}' 2>/dev/null | true", "delta": "0:00:00.012714", "end": "2016-10-18 15:40:21.770477", "rc": 0, "start": "2016-10-18 15:40:21.757763", "stderr": "", "stdout": "", "stdout_lines": [], "warnings": [] } } ``` see the part `cmd` of that command. This is clearly wrong. It is also not just a debugging issue, because the correct command would have a correct result. This one just returns nothing.
main
shell command not escaping double quotes correctly issue type bug report component name shell module ansible version ansible detached head last updated gmt os environment debian linux smp debian gnu linux summary for some formatting reasons double curly braces are needed in a shell command this worked in the past but how a hashtag is somehow produces by that workaround see the steps to reproduce steps to reproduce name get version of current docker engine shell usr bin docker version format client version dev null true register installed docker version debug var installed docker version when installed docker version is defined expected results the actual version of docker actual results the result is ok installed docker version changed false cmd usr bin docker version format client version dev null true delta end rc start stderr stdout stdout lines warnings see the part cmd of that command this is clearly wrong it is also not just a debugging issue because the correct command would have a correct result this one just returns nothing
1
983
2,594,403,638
IssuesEvent
2015-02-20 02:54:59
BALL-Project/ball
https://api.github.com/repos/BALL-Project/ball
closed
Model options can only be changed once
C: VIEW P: major R: fixed T: defect
**Reported by pebert on 30 Jun 41520711 22:10 UTC** BALLView 1.4.0 (June 29, 2011) QT 4.7.2 Ubuntu x64 - load molecule -> create representation (e.g. cartoon) - open model options -> change tube radius -> apply -> close window - open model options again -> try to change tube radius or restore defaults -> buttons ok/apply remain inactive - cannot be circumvented by deleting representation and recreating it (i.e. ok/apply remain inactive)
1.0
Model options can only be changed once - **Reported by pebert on 30 Jun 41520711 22:10 UTC** BALLView 1.4.0 (June 29, 2011) QT 4.7.2 Ubuntu x64 - load molecule -> create representation (e.g. cartoon) - open model options -> change tube radius -> apply -> close window - open model options again -> try to change tube radius or restore defaults -> buttons ok/apply remain inactive - cannot be circumvented by deleting representation and recreating it (i.e. ok/apply remain inactive)
non_main
model options can only be changed once reported by pebert on jun utc ballview june qt ubuntu load molecule create representation e g cartoon open model options change tube radius apply close window open model options again try to change tube radius or restore defaults buttons ok apply remain inactive cannot be circumvented by deleting representation and recreating it i e ok apply remain inactive
0
5,117
26,047,244,694
IssuesEvent
2022-12-22 15:22:42
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Double check the joining logic
type: bug work: backend status: draft restricted: maintainers
## Description * We need to come with a set of scenarios and the result we expect to achieve and double check our joining logic to ensure that the result matches our expectations. * Some critical scenarios: - Joining multiple times to the same table using different paths - Joining multiple times to the same table with the same path - Both the above cases for links to & from base table - Joining self referential tables with themselves by both forward & reverse links - Joining tables that link to base_table and contain circular references with multiple tables involved - The list is to be updated
True
Double check the joining logic - ## Description * We need to come with a set of scenarios and the result we expect to achieve and double check our joining logic to ensure that the result matches our expectations. * Some critical scenarios: - Joining multiple times to the same table using different paths - Joining multiple times to the same table with the same path - Both the above cases for links to & from base table - Joining self referential tables with themselves by both forward & reverse links - Joining tables that link to base_table and contain circular references with multiple tables involved - The list is to be updated
main
double check the joining logic description we need to come with a set of scenarios and the result we expect to achieve and double check our joining logic to ensure that the result matches our expectations some critical scenarios joining multiple times to the same table using different paths joining multiple times to the same table with the same path both the above cases for links to from base table joining self referential tables with themselves by both forward reverse links joining tables that link to base table and contain circular references with multiple tables involved the list is to be updated
1
5,306
26,800,970,858
IssuesEvent
2023-02-01 15:05:34
makubacki/mu_devops
https://api.github.com/repos/makubacki/mu_devops
closed
[Bug]: Test 3
state:needs-triage state:needs-owner state:needs-maintainer-feedback type:bug urgency:high
### Is there an existing issue for this? - [X] I have searched existing issues ### Current Behavior Test ### Expected Behavior Test ### Steps To Reproduce Test ### Build Environment ```markdown - OS(s): Test - Tool Chain(s): Test - Targets Impacted: Test ``` ### Version Information ```text Test ``` ### Urgency High ### Are you going to fix this? Someone else needs to fix it ### Do you need maintainer feedback? Maintainer feedback requested ### Anything else? _No response_
True
[Bug]: Test 3 - ### Is there an existing issue for this? - [X] I have searched existing issues ### Current Behavior Test ### Expected Behavior Test ### Steps To Reproduce Test ### Build Environment ```markdown - OS(s): Test - Tool Chain(s): Test - Targets Impacted: Test ``` ### Version Information ```text Test ``` ### Urgency High ### Are you going to fix this? Someone else needs to fix it ### Do you need maintainer feedback? Maintainer feedback requested ### Anything else? _No response_
main
test is there an existing issue for this i have searched existing issues current behavior test expected behavior test steps to reproduce test build environment markdown os s test tool chain s test targets impacted test version information text test urgency high are you going to fix this someone else needs to fix it do you need maintainer feedback maintainer feedback requested anything else no response
1
1,904
6,577,556,339
IssuesEvent
2017-09-12 01:44:22
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2_eip module returns true even when EIP is already assigned
affects_2.0 aws bug_report cloud waiting_on_maintainer
##### Issue Type: - Bug Report ##### Ansible Version: ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### Ansible Configuration: ##### Environment: Control machine: macosx el capitan Target machine: AWS RHEL7 instance - t2.large ##### Summary: An instance is already running with the EIP address. I then try to assign the same EIP to a new instance, and the assignment appears to succeed (though the instance comes up with no public ip address at all). ##### Steps To Reproduce: 1. Start an instance on AWS with an assigned EIP 2. Then start another instance on AWS and use the ec2_eip module and attempt to assign THE SAME EIP to it. 3. It should fail, but doesn't. ``` - name: associate an elastic IP with an instance ec2_eip: device_id={{ dev_id }} public_ip=52.36.100.30 region=us-west-2 register: instance ``` ##### Expected Results: It should fail. If I'm wrong and this is the expected result, then the documentation should be altered to make this VERY CLEAR. ##### Actual Results: ``` TASK [associate an elastic IP with an instance] ******************************** task path: /Users/fred/myscript.yml:90 ESTABLISH LOCAL CONNECTION FOR USER: fred 127.0.0.1 EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779 `" )' 127.0.0.1 PUT /var/folders/lg/c_x9gpnj3kx6ssg58nkdd2m00000gn/T/tmpQalVtR TO /Users/fred/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779/ec2_eip 127.0.0.1 EXEC /bin/sh -c 'LANG=en_US.US-ASCII LC_ALL=en_US.US-ASCII LC_MESSAGES=en_US.US-ASCII /usr/bin/python /Users/fred/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779/ec2_eip' changed: [localhost] => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "device_id": "i-42bef698", "ec2_url": null, "in_vpc": false, "profile": null, "public_ip": "52.36.100.30", "region": "us-west-2", "release_on_disassociation": false, "reuse_existing_ip_allowed": false, "security_token": null, "state": "present", "validate_certs": true, "wait_timeout": 300}, "module_name": "ec2_eip"}, "public_ip": "52.36.100.30"} ```
True
ec2_eip module returns true even when EIP is already assigned - ##### Issue Type: - Bug Report ##### Ansible Version: ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### Ansible Configuration: ##### Environment: Control machine: macosx el capitan Target machine: AWS RHEL7 instance - t2.large ##### Summary: An instance is already running with the EIP address. I then try to assign the same EIP to a new instance, and the assignment appears to succeed (though the instance comes up with no public ip address at all). ##### Steps To Reproduce: 1. Start an instance on AWS with an assigned EIP 2. Then start another instance on AWS and use the ec2_eip module and attempt to assign THE SAME EIP to it. 3. It should fail, but doesn't. ``` - name: associate an elastic IP with an instance ec2_eip: device_id={{ dev_id }} public_ip=52.36.100.30 region=us-west-2 register: instance ``` ##### Expected Results: It should fail. If I'm wrong and this is the expected result, then the documentation should be altered to make this VERY CLEAR. ##### Actual Results: ``` TASK [associate an elastic IP with an instance] ******************************** task path: /Users/fred/myscript.yml:90 ESTABLISH LOCAL CONNECTION FOR USER: fred 127.0.0.1 EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779 `" )' 127.0.0.1 PUT /var/folders/lg/c_x9gpnj3kx6ssg58nkdd2m00000gn/T/tmpQalVtR TO /Users/fred/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779/ec2_eip 127.0.0.1 EXEC /bin/sh -c 'LANG=en_US.US-ASCII LC_ALL=en_US.US-ASCII LC_MESSAGES=en_US.US-ASCII /usr/bin/python /Users/fred/.ansible/tmp/ansible-tmp-1456512356.09-272760837751779/ec2_eip' changed: [localhost] => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "device_id": "i-42bef698", "ec2_url": null, "in_vpc": false, "profile": null, "public_ip": "52.36.100.30", "region": "us-west-2", "release_on_disassociation": false, "reuse_existing_ip_allowed": false, "security_token": null, "state": "present", "validate_certs": true, "wait_timeout": 300}, "module_name": "ec2_eip"}, "public_ip": "52.36.100.30"} ```
main
eip module returns true even when eip is already assigned issue type bug report ansible version ansible config file configured module search path default w o overrides ansible configuration environment control machine macosx el capitan target machine aws instance large summary an instance is already running with the eip address i then try to assign the same eip to a new instance and the assignment appears to succeed though the instance comes up with no public ip address at all steps to reproduce start an instance on aws with an assigned eip then start another instance on aws and use the eip module and attempt to assign the same eip to it it should fail but doesn t name associate an elastic ip with an instance eip device id dev id public ip region us west register instance expected results it should fail if i m wrong and this is the expected result then the documentation should be altered to make this very clear actual results task task path users fred myscript yml establish local connection for user fred exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp put var folders lg c t tmpqalvtr to users fred ansible tmp ansible tmp eip exec bin sh c lang en us us ascii lc all en us us ascii lc messages en us us ascii usr bin python users fred ansible tmp ansible tmp eip changed changed true invocation module args aws access key null aws secret key null device id i url null in vpc false profile null public ip region us west release on disassociation false reuse existing ip allowed false security token null state present validate certs true wait timeout module name eip public ip
1
179,262
21,554,391,150
IssuesEvent
2022-04-30 06:33:58
Trinadh465/external_sonivox_AOSP10_r33
https://api.github.com/repos/Trinadh465/external_sonivox_AOSP10_r33
opened
CVE-2020-0385 (Medium) detected in sonivoxandroid-10.0.0_r41, sonivoxandroid-10.0.0_r41
security vulnerability
## CVE-2020-0385 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>sonivoxandroid-10.0.0_r41</b>, <b>sonivoxandroid-10.0.0_r41</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Parse_insh of eas_mdls.c, there is a possible out of bounds write due to an incorrect bounds check. This could lead to remote information disclosure in the media extractor with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-9 Android-10 Android-11 Android-8.0 Android-8.1Android ID: A-150160041 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0385>CVE-2020-0385</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-8.0.0_r50,android-8.1.0_r80,android-9.0.0_r60,android-10.0.0_r46</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-0385 (Medium) detected in sonivoxandroid-10.0.0_r41, sonivoxandroid-10.0.0_r41 - ## CVE-2020-0385 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>sonivoxandroid-10.0.0_r41</b>, <b>sonivoxandroid-10.0.0_r41</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Parse_insh of eas_mdls.c, there is a possible out of bounds write due to an incorrect bounds check. This could lead to remote information disclosure in the media extractor with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-9 Android-10 Android-11 Android-8.0 Android-8.1Android ID: A-150160041 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0385>CVE-2020-0385</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-8.0.0_r50,android-8.1.0_r80,android-9.0.0_r60,android-10.0.0_r46</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in sonivoxandroid sonivoxandroid cve medium severity vulnerability vulnerable libraries sonivoxandroid sonivoxandroid vulnerability details in parse insh of eas mdls c there is a possible out of bounds write due to an incorrect bounds check this could lead to remote information disclosure in the media extractor with no additional execution privileges needed user interaction is needed for exploitation product androidversions android android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android android android step up your open source security game with whitesource
0
204,607
15,503,890,485
IssuesEvent
2021-03-11 13:42:08
Slimefun/Slimefun4
https://api.github.com/repos/Slimefun/Slimefun4
opened
Ancient Rune Enchantment can enchant Slimefun's head based items.
🎯 Needs testing 🐞 Bug Report
<!-- FILL IN THE FORM BELOW --> ## :round_pushpin: Description (REQUIRED) <!-- A clear and detailed description of what went wrong. --> <!-- The more information you can provide, the easier we can handle this problem. --> <!-- Start writing below this line --> Ancient Rune Enchantment can enchant Slimefun's head based items. (most broken with uranium and if somebody use it on head) ## :bookmark_tabs: Steps to reproduce the Issue (REQUIRED) <!-- Tell us the exact steps to reproduce this issue, the more detailed the easier we can reproduce it. --> <!-- Youtube Videos and Screenshots are recommended!!! --> <!-- Start writing below this line --> 1. Acquire Ancient Rune [Enchantment] 2. Acquire head based Sf items. 3. Drop them and see! https://user-images.githubusercontent.com/34732414/110796152-37f84200-82aa-11eb-957c-55e2c5ea8c45.mp4 ## :bulb: Expected behavior (REQUIRED) <!-- What were you expecting to happen? --> <!-- What do you think would have been the correct behaviour? --> <!-- Start writing below this line --> Ancient Rune should not enchant Sf based items ## :compass: Environment (REQUIRED) <!-- Any issue without the exact version numbers will be closed! --> <!-- "latest" IS NOT A VERSION NUMBER. --> <!-- We recommend running "/sf versions" and showing us a screenshot of that. --> <!-- Make sure that the screenshot covers the entire output of that command. --> <!-- If your issue is related to other plugins, make sure to include the versions of these plugins too! --> ![image](https://user-images.githubusercontent.com/34732414/110796133-329af780-82aa-11eb-849b-35d2371177e5.png)
1.0
Ancient Rune Enchantment can enchant Slimefun's head based items. - <!-- FILL IN THE FORM BELOW --> ## :round_pushpin: Description (REQUIRED) <!-- A clear and detailed description of what went wrong. --> <!-- The more information you can provide, the easier we can handle this problem. --> <!-- Start writing below this line --> Ancient Rune Enchantment can enchant Slimefun's head based items. (most broken with uranium and if somebody use it on head) ## :bookmark_tabs: Steps to reproduce the Issue (REQUIRED) <!-- Tell us the exact steps to reproduce this issue, the more detailed the easier we can reproduce it. --> <!-- Youtube Videos and Screenshots are recommended!!! --> <!-- Start writing below this line --> 1. Acquire Ancient Rune [Enchantment] 2. Acquire head based Sf items. 3. Drop them and see! https://user-images.githubusercontent.com/34732414/110796152-37f84200-82aa-11eb-957c-55e2c5ea8c45.mp4 ## :bulb: Expected behavior (REQUIRED) <!-- What were you expecting to happen? --> <!-- What do you think would have been the correct behaviour? --> <!-- Start writing below this line --> Ancient Rune should not enchant Sf based items ## :compass: Environment (REQUIRED) <!-- Any issue without the exact version numbers will be closed! --> <!-- "latest" IS NOT A VERSION NUMBER. --> <!-- We recommend running "/sf versions" and showing us a screenshot of that. --> <!-- Make sure that the screenshot covers the entire output of that command. --> <!-- If your issue is related to other plugins, make sure to include the versions of these plugins too! --> ![image](https://user-images.githubusercontent.com/34732414/110796133-329af780-82aa-11eb-849b-35d2371177e5.png)
non_main
ancient rune enchantment can enchant slimefun s head based items round pushpin description required ancient rune enchantment can enchant slimefun s head based items most broken with uranium and if somebody use it on head bookmark tabs steps to reproduce the issue required acquire ancient rune acquire head based sf items drop them and see bulb expected behavior required ancient rune should not enchant sf based items compass environment required
0
311,258
26,779,250,937
IssuesEvent
2023-01-31 19:41:08
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/group6/saved_search_embeddable·ts - dashboard app - group 1 dashboard saved search embeddable view action leads to a saved search
Team:Presentation failed-test
A test failed on a tracked branch ``` Error: timed out waiting for Discover app on screen at onFailure (test/common/services/retry/retry_for_truthy.ts:39:13) at retryForSuccess (test/common/services/retry/retry_for_success.ts:59:13) at retryForTruthy (test/common/services/retry/retry_for_truthy.ts:27:3) at RetryService.waitFor (test/common/services/retry/retry.ts:59:5) at DiscoverPageObject.waitForDiscoverAppOnScreen (test/functional/page_objects/discover_page.ts:570:5) at Context.<anonymous> (test/functional/apps/dashboard/group6/saved_search_embeddable.ts:93:7) at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/19009#0182432e-ad03-4cf9-9473-1bac1285f492) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/dashboard/group6/saved_search_embeddable·ts","test.name":"dashboard app - group 1 dashboard saved search embeddable view action leads to a saved search","test.failCount":4}} -->
1.0
Failing test: Chrome UI Functional Tests.test/functional/apps/dashboard/group6/saved_search_embeddable·ts - dashboard app - group 1 dashboard saved search embeddable view action leads to a saved search - A test failed on a tracked branch ``` Error: timed out waiting for Discover app on screen at onFailure (test/common/services/retry/retry_for_truthy.ts:39:13) at retryForSuccess (test/common/services/retry/retry_for_success.ts:59:13) at retryForTruthy (test/common/services/retry/retry_for_truthy.ts:27:3) at RetryService.waitFor (test/common/services/retry/retry.ts:59:5) at DiscoverPageObject.waitForDiscoverAppOnScreen (test/functional/page_objects/discover_page.ts:570:5) at Context.<anonymous> (test/functional/apps/dashboard/group6/saved_search_embeddable.ts:93:7) at Object.apply (node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/19009#0182432e-ad03-4cf9-9473-1bac1285f492) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome UI Functional Tests.test/functional/apps/dashboard/group6/saved_search_embeddable·ts","test.name":"dashboard app - group 1 dashboard saved search embeddable view action leads to a saved search","test.failCount":4}} -->
non_main
failing test chrome ui functional tests test functional apps dashboard saved search embeddable·ts dashboard app group dashboard saved search embeddable view action leads to a saved search a test failed on a tracked branch error timed out waiting for discover app on screen at onfailure test common services retry retry for truthy ts at retryforsuccess test common services retry retry for success ts at retryfortruthy test common services retry retry for truthy ts at retryservice waitfor test common services retry retry ts at discoverpageobject waitfordiscoverapponscreen test functional page objects discover page ts at context test functional apps dashboard saved search embeddable ts at object apply node modules kbn test target node functional test runner lib mocha wrap function js first failure
0
2,355
8,409,595,953
IssuesEvent
2018-10-12 07:55:08
Homebrew/homebrew-cask
https://api.github.com/repos/Homebrew/homebrew-cask
closed
brew cask upgrade loses Launchpad app settings
awaiting maintainer feedback
#### General troubleshooting steps - [x] I have retried my command with `--force` and the issue is still present. - [x] I have checked the instructions for [reporting bugs](https://github.com/Homebrew/homebrew-cask#reporting-bugs). - [x] I made doubly sure this is not a [checksum does not match](https://github.com/Homebrew/homebrew-cask/blob/master/doc/reporting_bugs/a_cask_fails_to_install.md#checksum-does-not-match-error) error. - [x] I ran `brew update-reset && brew update` and retried my command. - [x] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [x] I checked there are no [open issues](https://github.com/Homebrew/homebrew-cask/issues) for the same problem. - [x] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/Homebrew/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue Launchpad is configurable. That is: you can order your apps or move an app to a folder and it will stay there. When you update it, it stays there. But when you uninstall and install it again, it will lose Launchpad settings. Today I updated my casks and Transmit got updated. I believe, it was uninstalled and installed again, so it lost Launchpad settings. Can we do something to prevent this from happening? <!-- Please DO NOT delete the backticks. Only change the “{{replace this}}” text. --> #### Command that failed ``` brew cask upgrade --greedy ``` #### Output of command with `--force --verbose --debug` n/a #### Output of `brew cask doctor` ``` ==> Homebrew Version 1.7.6 ==> macOS 10.13.6 ==> SIP Enabled ==> Java 1.8.0_181 ==> Gatekeeper support Enabled ==> Homebrew Cask Install Location <NONE> ==> Homebrew Cask Staging Location /usr/local/Caskroom ==> Homebrew Cask Taps: /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask (4033 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-drivers (151 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/ruby-macho-2.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/plist-3.4.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/ /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/backports-3.11.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/activesupport-5.2.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/tzinfo-1.2.5/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/thread_safe-0.3.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/minitest-5.11.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/i18n-1.1.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/concurrent-ruby-1.0.5/lib /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/gems/2.3.0/gems/did_you_mean-1.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew ==> Environment Variables LC_ALL="en_US.UTF-8" PATH="/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/zsh" ``` #### Output of `brew tap` ``` homebrew/cask homebrew/cask-drivers homebrew/core ```
True
brew cask upgrade loses Launchpad app settings - #### General troubleshooting steps - [x] I have retried my command with `--force` and the issue is still present. - [x] I have checked the instructions for [reporting bugs](https://github.com/Homebrew/homebrew-cask#reporting-bugs). - [x] I made doubly sure this is not a [checksum does not match](https://github.com/Homebrew/homebrew-cask/blob/master/doc/reporting_bugs/a_cask_fails_to_install.md#checksum-does-not-match-error) error. - [x] I ran `brew update-reset && brew update` and retried my command. - [x] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [x] I checked there are no [open issues](https://github.com/Homebrew/homebrew-cask/issues) for the same problem. - [x] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/Homebrew/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue Launchpad is configurable. That is: you can order your apps or move an app to a folder and it will stay there. When you update it, it stays there. But when you uninstall and install it again, it will lose Launchpad settings. Today I updated my casks and Transmit got updated. I believe, it was uninstalled and installed again, so it lost Launchpad settings. Can we do something to prevent this from happening? <!-- Please DO NOT delete the backticks. Only change the “{{replace this}}” text. --> #### Command that failed ``` brew cask upgrade --greedy ``` #### Output of command with `--force --verbose --debug` n/a #### Output of `brew cask doctor` ``` ==> Homebrew Version 1.7.6 ==> macOS 10.13.6 ==> SIP Enabled ==> Java 1.8.0_181 ==> Gatekeeper support Enabled ==> Homebrew Cask Install Location <NONE> ==> Homebrew Cask Staging Location /usr/local/Caskroom ==> Homebrew Cask Taps: /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask (4033 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-drivers (151 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/ruby-macho-2.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/plist-3.4.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/ /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/backports-3.11.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/activesupport-5.2.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/tzinfo-1.2.5/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/thread_safe-0.3.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/minitest-5.11.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/i18n-1.1.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle-standalone/bundler/../ruby/2.3.0/gems/concurrent-ruby-1.0.5/lib /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/gems/2.3.0/gems/did_you_mean-1.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/site_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/vendor_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.7/lib/ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew ==> Environment Variables LC_ALL="en_US.UTF-8" PATH="/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/zsh" ``` #### Output of `brew tap` ``` homebrew/cask homebrew/cask-drivers homebrew/core ```
main
brew cask upgrade loses launchpad app settings general troubleshooting steps i have retried my command with force and the issue is still present i have checked the instructions for i made doubly sure this is not a error i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i checked there are no for the same problem i understand that description of issue launchpad is configurable that is you can order your apps or move an app to a folder and it will stay there when you update it it stays there but when you uninstall and install it again it will lose launchpad settings today i updated my casks and transmit got updated i believe it was uninstalled and installed again so it lost launchpad settings can we do something to prevent this from happening command that failed brew cask upgrade greedy output of command with force verbose debug n a output of brew cask doctor homebrew version macos sip enabled java gatekeeper support enabled homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask taps usr local homebrew library taps homebrew homebrew cask casks usr local homebrew library taps homebrew homebrew cask drivers casks contents of load path usr local homebrew library homebrew vendor bundle standalone bundler ruby gems ruby macho lib usr local homebrew library homebrew vendor bundle standalone bundler ruby gems plist lib usr local homebrew library homebrew vendor bundle standalone bundler usr local homebrew library homebrew vendor bundle standalone bundler ruby gems backports lib usr local homebrew library homebrew vendor bundle standalone bundler ruby gems activesupport lib usr local homebrew library homebrew vendor bundle standalone bundler ruby gems tzinfo lib usr local homebrew library homebrew vendor bundle standalone bundler ruby gems thread safe lib usr local homebrew library homebrew vendor bundle standalone bundler ruby gems minitest lib usr local homebrew library homebrew vendor bundle standalone bundler ruby gems lib usr local homebrew library homebrew vendor bundle standalone bundler ruby gems concurrent ruby lib usr local homebrew library homebrew vendor portable ruby lib ruby gems gems did you mean lib usr local homebrew library homebrew vendor portable ruby lib ruby site ruby usr local homebrew library homebrew vendor portable ruby lib ruby site ruby usr local homebrew library homebrew vendor portable ruby lib ruby site ruby universal usr local homebrew library homebrew vendor portable ruby lib ruby site ruby usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby universal usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby usr local homebrew library homebrew vendor portable ruby lib ruby usr local homebrew library homebrew vendor portable ruby lib ruby usr local homebrew library homebrew vendor portable ruby lib ruby universal usr local homebrew library homebrew environment variables lc all en us utf path usr bin bin usr sbin sbin usr local homebrew library homebrew shims scm shell bin zsh output of brew tap homebrew cask homebrew cask drivers homebrew core
1
4,278
15,935,775,146
IssuesEvent
2021-04-14 10:16:18
nf-core/tools
https://api.github.com/repos/nf-core/tools
closed
Make docker image for nf-core/tools
automation command line tools low-priority
We have a root level `Dockerfile` in the tools repo which automatically builds at Docker Hub but it's the `nfcore/base` one that is used as a base for pipeline images: https://hub.docker.com/repository/docker/nfcore/base It would be good to also write a Dockerfile that builds the nf-core/tools package into an image that people can use (yes I know, it should be on BioContainers too, but doing `docker run nfcore/tools` is nicer). * [ ] Build a `Dockerfile` for tools which has as small a footprint as possible * [ ] Organise `Dockerfiles` into directories so we can have both base and tools * [ ] Remove DockerHub automated builds * [ ] Write GitHub Actions builds + push to DockerHub * [ ] Write some documentation
1.0
Make docker image for nf-core/tools - We have a root level `Dockerfile` in the tools repo which automatically builds at Docker Hub but it's the `nfcore/base` one that is used as a base for pipeline images: https://hub.docker.com/repository/docker/nfcore/base It would be good to also write a Dockerfile that builds the nf-core/tools package into an image that people can use (yes I know, it should be on BioContainers too, but doing `docker run nfcore/tools` is nicer). * [ ] Build a `Dockerfile` for tools which has as small a footprint as possible * [ ] Organise `Dockerfiles` into directories so we can have both base and tools * [ ] Remove DockerHub automated builds * [ ] Write GitHub Actions builds + push to DockerHub * [ ] Write some documentation
non_main
make docker image for nf core tools we have a root level dockerfile in the tools repo which automatically builds at docker hub but it s the nfcore base one that is used as a base for pipeline images it would be good to also write a dockerfile that builds the nf core tools package into an image that people can use yes i know it should be on biocontainers too but doing docker run nfcore tools is nicer build a dockerfile for tools which has as small a footprint as possible organise dockerfiles into directories so we can have both base and tools remove dockerhub automated builds write github actions builds push to dockerhub write some documentation
0
2,130
7,272,993,998
IssuesEvent
2018-02-21 02:05:12
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
brew cask style error
awaiting maintainer feedback bug core
#### General troubleshooting steps - [x] I have retried my command with `--force` and the issue is still present. - [x] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [x] None of the templates was appropriate for my issue, or I’m not sure. - [x] I ran `brew update-reset && brew update` and retried my command. - [x] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [x] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue Issue discovered while trying to run @vitorgalvao's `cask-repair` (getting red dot error outputs). #### Output of your command with `--verbose --debug` ``` $ brew cask style --verbose --debug /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require': cannot load such file -- parser (LoadError) from /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/gems/rubocop-0.52.1/lib/rubocop.rb:3:in `<top (required)>' from /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require' from /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/gems/rubocop-0.52.1/bin/rubocop:6:in `<top (required)>' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/bin/rubocop:22:in `load' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/bin/rubocop:22:in `<main>' Error: style check failed /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/style.rb:15:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/abstract_command.rb:35:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:98:in `run_command' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:168:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:100:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `rescue in run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:156:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:100:in `<main>' ``` #### Output of `brew cask doctor` ``` $ brew cask doctor ==> Homebrew-Cask Version Homebrew-Cask 1.5.4-12-ge904983 caskroom/homebrew-cask (git revision 87e5af; last commit 2018-02-12) ==> macOS 10.13.2 ==> SIP Enabled ==> Java N/A ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3913 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-drivers (141 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1159 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (173 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-eid (10 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/gems/2.3.0/gems/did_you_mean-1.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/universal-darwin9.0 ==> Environment Variables LC_ALL="en_US.UTF-8" PATH="/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/bash" ``` What is interesting is that after I started experiencing this issue I tried to `brew rm ruby@2.3` (which is what I had installed) and `brew install ruby` which got me to ruby 2.5. However the `$LOAD_PATH` in homebrew still has a bunch of 2.3.3 paths... ``` $ which ruby /usr/local/Cellar/ruby/2.5.0/bin/ruby $ which gem /usr/local/Cellar/ruby/2.5.0/bin/gem $ ruby --version ruby 2.5.0p0 (2017-12-25 revision 61468) [x86_64-darwin17] $ ruby env Traceback (most recent call last): ruby: No such file or directory -- env (LoadError) ``` When I previously had a problem with `cask-repair` there was `gem` command for removing `rubocop` and `rubocop-cask` which I tried again, but they were already removed and the gems were gone from `~/.gem` and the subdirs. It was after this that I tried to upgrade ruby. Help appreciated in getting `brew cask style` (and by extension `cask-repair`) working again!
True
brew cask style error - #### General troubleshooting steps - [x] I have retried my command with `--force` and the issue is still present. - [x] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [x] None of the templates was appropriate for my issue, or I’m not sure. - [x] I ran `brew update-reset && brew update` and retried my command. - [x] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [x] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue Issue discovered while trying to run @vitorgalvao's `cask-repair` (getting red dot error outputs). #### Output of your command with `--verbose --debug` ``` $ brew cask style --verbose --debug /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require': cannot load such file -- parser (LoadError) from /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/gems/rubocop-0.52.1/lib/rubocop.rb:3:in `<top (required)>' from /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require' from /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:55:in `require' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/gems/rubocop-0.52.1/bin/rubocop:6:in `<top (required)>' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/bin/rubocop:22:in `load' from /usr/local/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.3.0/bin/rubocop:22:in `<main>' Error: style check failed /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/style.rb:15:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/abstract_command.rb:35:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:98:in `run_command' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:168:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:100:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `rescue in run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:156:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:100:in `<main>' ``` #### Output of `brew cask doctor` ``` $ brew cask doctor ==> Homebrew-Cask Version Homebrew-Cask 1.5.4-12-ge904983 caskroom/homebrew-cask (git revision 87e5af; last commit 2018-02-12) ==> macOS 10.13.2 ==> SIP Enabled ==> Java N/A ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3913 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-drivers (141 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1159 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (173 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-eid (10 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/gems/2.3.0/gems/did_you_mean-1.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/site_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby/2.3.0/universal-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/vendor_ruby /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/x86_64-darwin9.0 /usr/local/Homebrew/Library/Homebrew/vendor/portable-ruby/2.3.3/lib/ruby/2.3.0/universal-darwin9.0 ==> Environment Variables LC_ALL="en_US.UTF-8" PATH="/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/bash" ``` What is interesting is that after I started experiencing this issue I tried to `brew rm ruby@2.3` (which is what I had installed) and `brew install ruby` which got me to ruby 2.5. However the `$LOAD_PATH` in homebrew still has a bunch of 2.3.3 paths... ``` $ which ruby /usr/local/Cellar/ruby/2.5.0/bin/ruby $ which gem /usr/local/Cellar/ruby/2.5.0/bin/gem $ ruby --version ruby 2.5.0p0 (2017-12-25 revision 61468) [x86_64-darwin17] $ ruby env Traceback (most recent call last): ruby: No such file or directory -- env (LoadError) ``` When I previously had a problem with `cask-repair` there was `gem` command for removing `rubocop` and `rubocop-cask` which I tried again, but they were already removed and the gems were gone from `~/.gem` and the subdirs. It was after this that I tried to upgrade ruby. Help appreciated in getting `brew cask style` (and by extension `cask-repair`) working again!
main
brew cask style error general troubleshooting steps i have retried my command with force and the issue is still present i have checked the instructions for or before opening the issue none of the templates was appropriate for my issue or i’m not sure i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i understand that description of issue issue discovered while trying to run vitorgalvao s cask repair getting red dot error outputs output of your command with verbose debug brew cask style verbose debug usr local homebrew library homebrew vendor portable ruby lib ruby rubygems core ext kernel require rb in require cannot load such file parser loaderror from usr local homebrew library homebrew vendor portable ruby lib ruby rubygems core ext kernel require rb in require from usr local homebrew library homebrew vendor bundle ruby gems rubocop lib rubocop rb in from usr local homebrew library homebrew vendor portable ruby lib ruby rubygems core ext kernel require rb in require from usr local homebrew library homebrew vendor portable ruby lib ruby rubygems core ext kernel require rb in require from usr local homebrew library homebrew vendor bundle ruby gems rubocop bin rubocop in from usr local homebrew library homebrew vendor bundle ruby bin rubocop in load from usr local homebrew library homebrew vendor bundle ruby bin rubocop in error style check failed usr local homebrew library homebrew cask lib hbc cli style rb in run usr local homebrew library homebrew cask lib hbc cli abstract command rb in run usr local homebrew library homebrew cask lib hbc cli rb in run command usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in error kernel exit usr local homebrew library homebrew cask lib hbc cli rb in exit usr local homebrew library homebrew cask lib hbc cli rb in rescue in run usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in output of brew cask doctor brew cask doctor homebrew cask version homebrew cask caskroom homebrew cask git revision last commit macos sip enabled java n a homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask taps usr local homebrew library taps caskroom homebrew cask casks usr local homebrew library taps caskroom homebrew drivers casks usr local homebrew library taps caskroom homebrew fonts casks usr local homebrew library taps caskroom homebrew versions casks usr local homebrew library taps caskroom homebrew eid casks contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew usr local homebrew library homebrew vendor portable ruby lib ruby gems gems did you mean lib usr local homebrew library homebrew vendor portable ruby lib ruby site ruby usr local homebrew library homebrew vendor portable ruby lib ruby site ruby usr local homebrew library homebrew vendor portable ruby lib ruby site ruby universal usr local homebrew library homebrew vendor portable ruby lib ruby site ruby usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby universal usr local homebrew library homebrew vendor portable ruby lib ruby vendor ruby usr local homebrew library homebrew vendor portable ruby lib ruby usr local homebrew library homebrew vendor portable ruby lib ruby usr local homebrew library homebrew vendor portable ruby lib ruby universal environment variables lc all en us utf path usr bin bin usr sbin sbin usr local homebrew library homebrew shims scm shell bin bash what is interesting is that after i started experiencing this issue i tried to brew rm ruby which is what i had installed and brew install ruby which got me to ruby however the load path in homebrew still has a bunch of paths which ruby usr local cellar ruby bin ruby which gem usr local cellar ruby bin gem ruby version ruby revision ruby env traceback most recent call last ruby no such file or directory env loaderror when i previously had a problem with cask repair there was gem command for removing rubocop and rubocop cask which i tried again but they were already removed and the gems were gone from gem and the subdirs it was after this that i tried to upgrade ruby help appreciated in getting brew cask style and by extension cask repair working again
1
196,458
15,594,983,909
IssuesEvent
2021-03-18 14:27:00
rocksdanister/lively
https://api.github.com/repos/rocksdanister/lively
reopened
MS Edge web browser crash.
documentation
**Before proceeding** Check the Wiki: https://github.com/rocksdanister/lively/wiki/Common-Problems and Issues: https://github.com/rocksdanister/lively/issues?q=is%3Aopen+is%3Aissue to see if your problem is already addressed. **Describe the bug** A clear and concise description of what the bug is. **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots/Video** If applicable, add screenshots/video to help explain your problem. **Desktop (please complete the following information):** - OS: [e.g. Windows 10 1909] - Which wallpaper was running at the time if any. **Additional context** Add any other context about the problem here. **Log file (Important)** Share the diagnostic files by: Open Lively settings (Gear icon) -> Misc. -> Log File or Navigate to Appdata folder `C:\Users\<UserName>\AppData\Local\Lively Wallpaper\logs` [lively_log_20210318_111113.zip](https://github.com/rocksdanister/lively/files/6163177/lively_log_20210318_111113.zip) [lively_log_20210318_111113.zip](https://github.com/rocksdanister/lively/files/6163178/lively_log_20210318_111113.zip)
1.0
MS Edge web browser crash. - **Before proceeding** Check the Wiki: https://github.com/rocksdanister/lively/wiki/Common-Problems and Issues: https://github.com/rocksdanister/lively/issues?q=is%3Aopen+is%3Aissue to see if your problem is already addressed. **Describe the bug** A clear and concise description of what the bug is. **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots/Video** If applicable, add screenshots/video to help explain your problem. **Desktop (please complete the following information):** - OS: [e.g. Windows 10 1909] - Which wallpaper was running at the time if any. **Additional context** Add any other context about the problem here. **Log file (Important)** Share the diagnostic files by: Open Lively settings (Gear icon) -> Misc. -> Log File or Navigate to Appdata folder `C:\Users\<UserName>\AppData\Local\Lively Wallpaper\logs` [lively_log_20210318_111113.zip](https://github.com/rocksdanister/lively/files/6163177/lively_log_20210318_111113.zip) [lively_log_20210318_111113.zip](https://github.com/rocksdanister/lively/files/6163178/lively_log_20210318_111113.zip)
non_main
ms edge web browser crash before proceeding check the wiki and issues to see if your problem is already addressed describe the bug a clear and concise description of what the bug is to reproduce steps to reproduce the behavior go to click on scroll down to see error expected behavior a clear and concise description of what you expected to happen screenshots video if applicable add screenshots video to help explain your problem desktop please complete the following information os which wallpaper was running at the time if any additional context add any other context about the problem here log file important share the diagnostic files by open lively settings gear icon misc log file or navigate to appdata folder c users appdata local lively wallpaper logs
0
2,163
2,524,254,815
IssuesEvent
2015-01-20 16:37:08
NLeSC/PattyVis
https://api.github.com/repos/NLeSC/PattyVis
opened
Move $scope.$watch site update to core
Low Priority
For issue #42 we need the $scope.$watch that is watching the filtered sites list to be moved to core. We need that $watcher to call an updating function for both the minimap site-footprints as well as the highlighted SiteBoxes from issue #42. In any case, it may make more sense to put it there, since sites are a core feature, and the sites.service is there as well.
1.0
Move $scope.$watch site update to core - For issue #42 we need the $scope.$watch that is watching the filtered sites list to be moved to core. We need that $watcher to call an updating function for both the minimap site-footprints as well as the highlighted SiteBoxes from issue #42. In any case, it may make more sense to put it there, since sites are a core feature, and the sites.service is there as well.
non_main
move scope watch site update to core for issue we need the scope watch that is watching the filtered sites list to be moved to core we need that watcher to call an updating function for both the minimap site footprints as well as the highlighted siteboxes from issue in any case it may make more sense to put it there since sites are a core feature and the sites service is there as well
0
127,841
10,490,336,373
IssuesEvent
2019-09-25 08:48:04
voyages-sncf-technologies/hesperides
https://api.github.com/repos/voyages-sncf-technologies/hesperides
closed
BDD: cas de MAJ de propriété avec le même nom mais 2 valeurs différentes dans la payload
good first issue test-manquant
cf. https://github.com/voyages-sncf-technologies/hesperides/pull/647/files#diff-6
1.0
BDD: cas de MAJ de propriété avec le même nom mais 2 valeurs différentes dans la payload - cf. https://github.com/voyages-sncf-technologies/hesperides/pull/647/files#diff-6
non_main
bdd cas de maj de propriété avec le même nom mais valeurs différentes dans la payload cf
0
192,230
22,215,918,558
IssuesEvent
2022-06-08 01:37:03
AlexRogalskiy/github-action-node-dependency
https://api.github.com/repos/AlexRogalskiy/github-action-node-dependency
closed
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js - autoclosed
security vulnerability
## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: github-action-node-dependency/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: github-action-node-dependency/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-node-dependency/commit/856c28e2689dd54cb1335a960165ae66275b404e">856c28e2689dd54cb1335a960165ae66275b404e</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js - autoclosed - ## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: github-action-node-dependency/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: github-action-node-dependency/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-node-dependency/commit/856c28e2689dd54cb1335a960165ae66275b404e">856c28e2689dd54cb1335a960165ae66275b404e</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file github action node dependency node modules redeyed examples browser index html path to vulnerable library github action node dependency node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
33,698
7,198,579,508
IssuesEvent
2018-02-05 13:21:31
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
opened
Folders::Incorrect error message is displayed "Document name already exist" upon not entering document name
Defect P2 SumFive Team
Steps To Replicate : 1. Launch the url : 2. Login as HR Admin user 3. Navigate to 'Settings' menu 4. Select 'Folders' menu >> click on + symbol to add a new folder 5. Go to 'Documents' section displayed at the right side pane >> click on "Upload" 6. popup windiow is displayed with option "dropfiles here or click to upload" >> click on it 7. Browser a PDF document and without entering any document name click on save button Experienced Behavior : Observed that incorrect error message is displayed "Document name already exist" Expected Behavior :Ensure that insted of displaying invalid error message application should prompt the user to enter "Document name " which was left empty
1.0
Folders::Incorrect error message is displayed "Document name already exist" upon not entering document name - Steps To Replicate : 1. Launch the url : 2. Login as HR Admin user 3. Navigate to 'Settings' menu 4. Select 'Folders' menu >> click on + symbol to add a new folder 5. Go to 'Documents' section displayed at the right side pane >> click on "Upload" 6. popup windiow is displayed with option "dropfiles here or click to upload" >> click on it 7. Browser a PDF document and without entering any document name click on save button Experienced Behavior : Observed that incorrect error message is displayed "Document name already exist" Expected Behavior :Ensure that insted of displaying invalid error message application should prompt the user to enter "Document name " which was left empty
non_main
folders incorrect error message is displayed document name already exist upon not entering document name steps to replicate launch the url login as hr admin user navigate to settings menu select folders menu click on symbol to add a new folder go to documents section displayed at the right side pane click on upload popup windiow is displayed with option dropfiles here or click to upload click on it browser a pdf document and without entering any document name click on save button experienced behavior observed that incorrect error message is displayed document name already exist expected behavior ensure that insted of displaying invalid error message application should prompt the user to enter document name which was left empty
0
53,722
23,041,390,487
IssuesEvent
2022-07-23 07:30:26
amplication/amplication
https://api.github.com/repos/amplication/amplication
closed
Upgrade to Apollo Server 3
type: dependencies @amplication/server @amplication/data-service-generator status: accepted
### What happened? Apollo Server 3 introduced new graphql sandbox: https://www.apollographql.com/docs/apollo-server/testing/build-run-queries/ https://www.apollographql.com/docs/studio/explorer/explorer/#account-free-sandbox On this version, it is easier to build queries and understand the schemas and the types. According to nest.js documentation, we first need to upgrade the @nestjs/graphql on data-service-generator to version>=9 and we may add @nestjs/apollo^10 (which we are not using right now) https://docs.nestjs.com/graphql/quick-start#installation ### What you expected to happen I expect to get the new Apollo Server graphql sandbox ### How to reproduce Generate an app, follow the instruction: to run the environment (https://docs.amplication.com/docs/getting-started/dev-env), go to localhost:3000/graphql and see the old version of Apollo server playground ### Amplication version 0..11.4 ### Environment node version 14.17.0 npm version 7.3.0 macOS ### Are you willing to submit PR? Yes I am willing to submit a PR! - [x] #2394 - [x] #2395
1.0
Upgrade to Apollo Server 3 - ### What happened? Apollo Server 3 introduced new graphql sandbox: https://www.apollographql.com/docs/apollo-server/testing/build-run-queries/ https://www.apollographql.com/docs/studio/explorer/explorer/#account-free-sandbox On this version, it is easier to build queries and understand the schemas and the types. According to nest.js documentation, we first need to upgrade the @nestjs/graphql on data-service-generator to version>=9 and we may add @nestjs/apollo^10 (which we are not using right now) https://docs.nestjs.com/graphql/quick-start#installation ### What you expected to happen I expect to get the new Apollo Server graphql sandbox ### How to reproduce Generate an app, follow the instruction: to run the environment (https://docs.amplication.com/docs/getting-started/dev-env), go to localhost:3000/graphql and see the old version of Apollo server playground ### Amplication version 0..11.4 ### Environment node version 14.17.0 npm version 7.3.0 macOS ### Are you willing to submit PR? Yes I am willing to submit a PR! - [x] #2394 - [x] #2395
non_main
upgrade to apollo server what happened apollo server introduced new graphql sandbox on this version it is easier to build queries and understand the schemas and the types according to nest js documentation we first need to upgrade the nestjs graphql on data service generator to version and we may add nestjs apollo which we are not using right now what you expected to happen i expect to get the new apollo server graphql sandbox how to reproduce generate an app follow the instruction to run the environment go to localhost graphql and see the old version of apollo server playground amplication version environment node version npm version macos are you willing to submit pr yes i am willing to submit a pr
0
3,310
12,820,294,593
IssuesEvent
2020-07-06 05:23:29
geolexica/geolexica-server
https://api.github.com/repos/geolexica/geolexica-server
closed
Configuration is not inherited properly
bug maintainability
The configuration written to `_config.yml` in this gem should be mixed into sites' configuration. But apparently it is not, that's why https://github.com/geolexica/osgeo.geolexica.org/commit/53c87bde71152250e6591fe29242285bb8f1b2dc was required. Maybe Jekyll upgrade will fix it.
True
Configuration is not inherited properly - The configuration written to `_config.yml` in this gem should be mixed into sites' configuration. But apparently it is not, that's why https://github.com/geolexica/osgeo.geolexica.org/commit/53c87bde71152250e6591fe29242285bb8f1b2dc was required. Maybe Jekyll upgrade will fix it.
main
configuration is not inherited properly the configuration written to config yml in this gem should be mixed into sites configuration but apparently it is not that s why was required maybe jekyll upgrade will fix it
1
1,765
6,575,022,287
IssuesEvent
2017-09-11 14:48:15
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Timeout: Cannot save running-config with nxos_command or save: yes in nxos_config
affects_2.3 bug_report networking waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME nxos_command nxos_config ##### ANSIBLE VERSION ``` ansible 2.3.0~git20161010.03765ba config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION inventory = ./hosts gathering = explicit roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles private_role_vars = yes log_path = /var/log/ansible.log fact_caching = redis fact_caching_timeout = 86400 retry_files_enabled = False ##### OS / ENVIRONMENT - host: Ubuntu 16.04 4.4.0 - target: **NX-OSv 7.3(0)D1(1)** ##### SUMMARY cf. title ##### STEPS TO REPRODUCE **Inventory ./hosts**: ``` [all:vars] nms_mgt_ip_address=172.21.100.1 [spines] NX_OSv_Spine_11 ansible_host=172.21.100.11 NX_OSv_Spine_12 ansible_host=172.21.100.12 ``` Structure passed as "**provider**": connections.ssh Defined in group_vars/nx_osv/connections.yml and a symbolic link in roles/nxos_snmp/defaults points to nx_osv ``` connections ... nxapi: transport: nxapi host: "{{ ansible_host }}" # ansible_port port: "{{ http.port }}" # ansible_user username: admin password: xxxxxxxx # enable_secret_password auth_pass: xxxxxxxx # http or https use_ssl: no validate_certs: "{{ https.validate_certs }}" ``` **Role**: nxos_snmp: ``` - include_vars: "../defaults/{{ os_family }}/connections.yml" ... - name: Saving the running-config nxos_command: provider: "{{ connections.nxapi }}" commands: - "copy running-config startup-config" register: result ``` **Playbook**: ``` - name: Configuring SNMP on NX-OS/NX-OSv hosts: - nx_osv roles: - nxos_snmp ``` ##### EXPECTED RESULTS We should be able to save the modifications. ##### ACTUAL RESULTS ``` TASK [nxos_snmp : Saving the running-config] *********************************** fatal: [NX_OSv_Spine_12]: FAILED! => {"changed": false, "clierror": "Syntax error while parsing 'copy running-config startup-config | xml '\n\n\nCmd exec error.\n", "code": "400", "failed": true, "input": "copy running-config startup-config", "msg": "Input CLI command error", "output": {"clierror": "Syntax error while parsing 'copy running-config startup-config | xml '\n\n\nCmd exec error.\n", "code": "400", "input": "copy running-config startup-config", "msg": "Input CLI command error"}, "url": "http://172.21.100.12:8080/ins"} ``` No issue when configuring through the CLI: ``` NX_OSv_Spine_12# copy running-config startup-config [########################################] 100% Copy complete. NX_OSv_Spine_12# show startup-config !Command: show startup-config !Time: Mon Oct 10 18:12:08 2016 !Startup config saved at: Mon Oct 10 18:01:55 2016 version 7.3(0)D1(1) power redundancy-mode redundant ```
True
Timeout: Cannot save running-config with nxos_command or save: yes in nxos_config - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME nxos_command nxos_config ##### ANSIBLE VERSION ``` ansible 2.3.0~git20161010.03765ba config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION inventory = ./hosts gathering = explicit roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles private_role_vars = yes log_path = /var/log/ansible.log fact_caching = redis fact_caching_timeout = 86400 retry_files_enabled = False ##### OS / ENVIRONMENT - host: Ubuntu 16.04 4.4.0 - target: **NX-OSv 7.3(0)D1(1)** ##### SUMMARY cf. title ##### STEPS TO REPRODUCE **Inventory ./hosts**: ``` [all:vars] nms_mgt_ip_address=172.21.100.1 [spines] NX_OSv_Spine_11 ansible_host=172.21.100.11 NX_OSv_Spine_12 ansible_host=172.21.100.12 ``` Structure passed as "**provider**": connections.ssh Defined in group_vars/nx_osv/connections.yml and a symbolic link in roles/nxos_snmp/defaults points to nx_osv ``` connections ... nxapi: transport: nxapi host: "{{ ansible_host }}" # ansible_port port: "{{ http.port }}" # ansible_user username: admin password: xxxxxxxx # enable_secret_password auth_pass: xxxxxxxx # http or https use_ssl: no validate_certs: "{{ https.validate_certs }}" ``` **Role**: nxos_snmp: ``` - include_vars: "../defaults/{{ os_family }}/connections.yml" ... - name: Saving the running-config nxos_command: provider: "{{ connections.nxapi }}" commands: - "copy running-config startup-config" register: result ``` **Playbook**: ``` - name: Configuring SNMP on NX-OS/NX-OSv hosts: - nx_osv roles: - nxos_snmp ``` ##### EXPECTED RESULTS We should be able to save the modifications. ##### ACTUAL RESULTS ``` TASK [nxos_snmp : Saving the running-config] *********************************** fatal: [NX_OSv_Spine_12]: FAILED! => {"changed": false, "clierror": "Syntax error while parsing 'copy running-config startup-config | xml '\n\n\nCmd exec error.\n", "code": "400", "failed": true, "input": "copy running-config startup-config", "msg": "Input CLI command error", "output": {"clierror": "Syntax error while parsing 'copy running-config startup-config | xml '\n\n\nCmd exec error.\n", "code": "400", "input": "copy running-config startup-config", "msg": "Input CLI command error"}, "url": "http://172.21.100.12:8080/ins"} ``` No issue when configuring through the CLI: ``` NX_OSv_Spine_12# copy running-config startup-config [########################################] 100% Copy complete. NX_OSv_Spine_12# show startup-config !Command: show startup-config !Time: Mon Oct 10 18:12:08 2016 !Startup config saved at: Mon Oct 10 18:01:55 2016 version 7.3(0)D1(1) power redundancy-mode redundant ```
main
timeout cannot save running config with nxos command or save yes in nxos config issue type bug report component name nxos command nxos config ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration inventory hosts gathering explicit roles path home actionmystique program files ubuntu ansible roles roles private role vars yes log path var log ansible log fact caching redis fact caching timeout retry files enabled false os environment host ubuntu target nx osv summary cf title steps to reproduce inventory hosts nms mgt ip address nx osv spine ansible host nx osv spine ansible host structure passed as provider connections ssh defined in group vars nx osv connections yml and a symbolic link in roles nxos snmp defaults points to nx osv connections nxapi transport nxapi host ansible host ansible port port http port ansible user username admin password xxxxxxxx enable secret password auth pass xxxxxxxx http or https use ssl no validate certs https validate certs role nxos snmp include vars defaults os family connections yml name saving the running config nxos command provider connections nxapi commands copy running config startup config register result playbook name configuring snmp on nx os nx osv hosts nx osv roles nxos snmp expected results we should be able to save the modifications actual results task fatal failed changed false clierror syntax error while parsing copy running config startup config xml n n ncmd exec error n code failed true input copy running config startup config msg input cli command error output clierror syntax error while parsing copy running config startup config xml n n ncmd exec error n code input copy running config startup config msg input cli command error url no issue when configuring through the cli nx osv spine copy running config startup config copy complete nx osv spine show startup config command show startup config time mon oct startup config saved at mon oct version power redundancy mode redundant
1
2,872
10,276,018,781
IssuesEvent
2019-08-24 13:38:19
arcticicestudio/arctic
https://api.github.com/repos/arcticicestudio/arctic
opened
lint-staged
context-workflow scope-dx scope-maintainability scope-quality type-feature
<p align="center"><img src="s://user-images.githubusercontent.com/7836623/63638143-c84d4280-c684-11e9-93cf-98662c6c0168.png" width="25%" /></p> Integrate [lint-staged][gh-lint-staged] to run linters against staged Git files to prevent to add code that violates any style guide into the code base. <p align="center"><img src="https://user-images.githubusercontent.com/7836623/63638144-c84d4280-c684-11e9-8ba1-1cec576a8fdb.gif" width="80%" /></p> ### Configuration The configuration file `lint-staged.config.js` will be placed in the project root and includes the command that should be run for matching file extensions (globs). It will include at least the three following entries with the same order as listed here: 1. `prettier --list-different` - Run Prettier (#32) against `*.{js,json,md,mdx,ts,tsx,yml}` to ensure all files are formatted correctly. The `--list-different` prints the found files that are not conform to the Prettier configuration. 2. `eslint` - Run ESLint (#30) against `*.{js}` to ensure all JavaScript files are compliant to the style guide after being formatted with Prettier. 3. `eslint` - Run ESLint (#30) against `*.{ts,tsx}` to ensure all JavaScript files are compliant to the style guide after being formatted with Prettier. 4. `remark --no-stdout` - Run remark-lint (#27) against `*.md` to ensure all Markdown files are compliant to the style guide. The `--no-stdout` flag suppresses the output of the parsed file content. ## Tasks - [ ] Install [lint-staged][npm-lint-staged] package. - [ ] Implement `lint-staged.config.js` configuration file. [gh-lint-staged]: https://github.com/okonet/lint-staged [npm-lint-staged]: https://www.npmjs.com/package/lint-staged
True
lint-staged - <p align="center"><img src="s://user-images.githubusercontent.com/7836623/63638143-c84d4280-c684-11e9-93cf-98662c6c0168.png" width="25%" /></p> Integrate [lint-staged][gh-lint-staged] to run linters against staged Git files to prevent to add code that violates any style guide into the code base. <p align="center"><img src="https://user-images.githubusercontent.com/7836623/63638144-c84d4280-c684-11e9-8ba1-1cec576a8fdb.gif" width="80%" /></p> ### Configuration The configuration file `lint-staged.config.js` will be placed in the project root and includes the command that should be run for matching file extensions (globs). It will include at least the three following entries with the same order as listed here: 1. `prettier --list-different` - Run Prettier (#32) against `*.{js,json,md,mdx,ts,tsx,yml}` to ensure all files are formatted correctly. The `--list-different` prints the found files that are not conform to the Prettier configuration. 2. `eslint` - Run ESLint (#30) against `*.{js}` to ensure all JavaScript files are compliant to the style guide after being formatted with Prettier. 3. `eslint` - Run ESLint (#30) against `*.{ts,tsx}` to ensure all JavaScript files are compliant to the style guide after being formatted with Prettier. 4. `remark --no-stdout` - Run remark-lint (#27) against `*.md` to ensure all Markdown files are compliant to the style guide. The `--no-stdout` flag suppresses the output of the parsed file content. ## Tasks - [ ] Install [lint-staged][npm-lint-staged] package. - [ ] Implement `lint-staged.config.js` configuration file. [gh-lint-staged]: https://github.com/okonet/lint-staged [npm-lint-staged]: https://www.npmjs.com/package/lint-staged
main
lint staged integrate to run linters against staged git files to prevent to add code that violates any style guide into the code base configuration the configuration file lint staged config js will be placed in the project root and includes the command that should be run for matching file extensions globs it will include at least the three following entries with the same order as listed here prettier list different run prettier against js json md mdx ts tsx yml to ensure all files are formatted correctly the list different prints the found files that are not conform to the prettier configuration eslint run eslint against js to ensure all javascript files are compliant to the style guide after being formatted with prettier eslint run eslint against ts tsx to ensure all javascript files are compliant to the style guide after being formatted with prettier remark no stdout run remark lint against md to ensure all markdown files are compliant to the style guide the no stdout flag suppresses the output of the parsed file content tasks install package implement lint staged config js configuration file
1
1,142
4,999,054,022
IssuesEvent
2016-12-09 21:55:55
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
git: Add an extra_args option to the git module.
affects_2.0 feature_idea waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME git module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = /home/duncan/dev/agile/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY I would like to set extra git options, but still use the ansible `git` module, rather than having to use `shell/command` instead. ##### STEPS TO REPRODUCE Specifically, I would like to be able to add a proxy config to a particular git clone. This could work in the same way as the `pip` module - for example: ``` yaml - name: Cloning repo git: repo: {{ git_repo }} version: 'dev' dest: {{ repo_dir }} extra_args: '--config "http.proxy=proxyHost:proxyPort"' ``` but there are lots of other arguments that one might want to pass.
True
git: Add an extra_args option to the git module. - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME git module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = /home/duncan/dev/agile/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY I would like to set extra git options, but still use the ansible `git` module, rather than having to use `shell/command` instead. ##### STEPS TO REPRODUCE Specifically, I would like to be able to add a proxy config to a particular git clone. This could work in the same way as the `pip` module - for example: ``` yaml - name: Cloning repo git: repo: {{ git_repo }} version: 'dev' dest: {{ repo_dir }} extra_args: '--config "http.proxy=proxyHost:proxyPort"' ``` but there are lots of other arguments that one might want to pass.
main
git add an extra args option to the git module issue type feature idea component name git module ansible version ansible config file home duncan dev agile ansible cfg configured module search path default w o overrides configuration n a os environment n a summary i would like to set extra git options but still use the ansible git module rather than having to use shell command instead steps to reproduce specifically i would like to be able to add a proxy config to a particular git clone this could work in the same way as the pip module for example yaml name cloning repo git repo git repo version dev dest repo dir extra args config http proxy proxyhost proxyport but there are lots of other arguments that one might want to pass
1
5,423
27,211,281,391
IssuesEvent
2023-02-20 16:44:59
centerofci/mathesar-website
https://api.github.com/repos/centerofci/mathesar-website
closed
Update website CTAs to promote SaaS version and email list
restricted: maintainers status: ready type: enhancement work: product
Myra sent us several suggestions [here](https://docs.google.com/document/d/1tNFPelG8pMMN0zPH_XvbmE-e9o9v6qzxZkOXKibXxcI/edit). This issue will implement all the website related ones. - [x] Homepage - 4 boxes changes - [x] Developer How it Works - 4 boxes changes - [x] Mailing list CTA - [x] add checkboxes to all pages - [x] update header and copy - [x] Ensure checkboxes propagate to MailerLite - @kgodey - [x] Non-Technical How it Works - [x] update CTA buttons - [x] update mailing list CTA header - [x] About Us page - [x] update sustainability copy - @kgodey - [x] FAQ page - [x] add new FAQ
True
Update website CTAs to promote SaaS version and email list - Myra sent us several suggestions [here](https://docs.google.com/document/d/1tNFPelG8pMMN0zPH_XvbmE-e9o9v6qzxZkOXKibXxcI/edit). This issue will implement all the website related ones. - [x] Homepage - 4 boxes changes - [x] Developer How it Works - 4 boxes changes - [x] Mailing list CTA - [x] add checkboxes to all pages - [x] update header and copy - [x] Ensure checkboxes propagate to MailerLite - @kgodey - [x] Non-Technical How it Works - [x] update CTA buttons - [x] update mailing list CTA header - [x] About Us page - [x] update sustainability copy - @kgodey - [x] FAQ page - [x] add new FAQ
main
update website ctas to promote saas version and email list myra sent us several suggestions this issue will implement all the website related ones homepage boxes changes developer how it works boxes changes mailing list cta add checkboxes to all pages update header and copy ensure checkboxes propagate to mailerlite kgodey non technical how it works update cta buttons update mailing list cta header about us page update sustainability copy kgodey faq page add new faq
1
92,175
18,785,609,111
IssuesEvent
2021-11-08 11:48:22
hashicorp/terraform-ls
https://api.github.com/repos/hashicorp/terraform-ls
opened
Report mismatching types of variables in tfvars
enhancement textDocument/codeAction textDocument/publishDiagnostics
### Use-cases Users with more complex modules and many variables may not always notice mismatching types within variable files. For example, say we have the following variable declaration ```hcl variable "example" { type = list(string) } ``` and then a `terraform.tfvars` file entry: ```hcl example = "foobar" ``` User may not immediately spot the mismatching type here, if they have many variables and variables of more complex types. ### Attempted Solutions Manual inspection or `terraform validate`. ### Proposal Report mismatching types of variables in any `*.tfvars` and `*.tfvars.json` via `textDocument/publishDiagnostics`.
1.0
Report mismatching types of variables in tfvars - ### Use-cases Users with more complex modules and many variables may not always notice mismatching types within variable files. For example, say we have the following variable declaration ```hcl variable "example" { type = list(string) } ``` and then a `terraform.tfvars` file entry: ```hcl example = "foobar" ``` User may not immediately spot the mismatching type here, if they have many variables and variables of more complex types. ### Attempted Solutions Manual inspection or `terraform validate`. ### Proposal Report mismatching types of variables in any `*.tfvars` and `*.tfvars.json` via `textDocument/publishDiagnostics`.
non_main
report mismatching types of variables in tfvars use cases users with more complex modules and many variables may not always notice mismatching types within variable files for example say we have the following variable declaration hcl variable example type list string and then a terraform tfvars file entry hcl example foobar user may not immediately spot the mismatching type here if they have many variables and variables of more complex types attempted solutions manual inspection or terraform validate proposal report mismatching types of variables in any tfvars and tfvars json via textdocument publishdiagnostics
0
26,663
7,857,035,941
IssuesEvent
2018-06-21 09:29:52
fossasia/susi_skill_cms
https://api.github.com/repos/fossasia/susi_skill_cms
opened
Bot Name should be same as the Skill Name
Botbuilder enhancement
**Actual Behaviour** Bot Name and Skill Name are separate fields. <!-- Please state here what is currently happening. --> **Expected Behaviour** Bot Name should be same as the Skill Name. They should be a single field. <!-- State here what the feature should enable the user to do. --> **Would you like to work on the issue?** Yes <!-- Please let us know if you can work on it or the issue should be assigned to someone else. -->
1.0
Bot Name should be same as the Skill Name - **Actual Behaviour** Bot Name and Skill Name are separate fields. <!-- Please state here what is currently happening. --> **Expected Behaviour** Bot Name should be same as the Skill Name. They should be a single field. <!-- State here what the feature should enable the user to do. --> **Would you like to work on the issue?** Yes <!-- Please let us know if you can work on it or the issue should be assigned to someone else. -->
non_main
bot name should be same as the skill name actual behaviour bot name and skill name are separate fields expected behaviour bot name should be same as the skill name they should be a single field would you like to work on the issue yes
0
144,125
22,281,813,020
IssuesEvent
2022-06-11 01:59:36
DeveloperAcademy-POSTECH/MC2-Team14-OXY
https://api.github.com/repos/DeveloperAcademy-POSTECH/MC2-Team14-OXY
closed
[Feature] 카드 long press & swipe 제스처
feature design 다니
## Description 카드를 꾹 눌러서 위/아래로 보내는 제스처 구현 ## ScreenShot ## To-do - [ ] 카드를 꾹 누르는 제스처 + 드래그 제스처 합치기 - [ ] 카드를 위로 보내기 - [ ] 카드를 아래로 내리기 ## Etc
1.0
[Feature] 카드 long press & swipe 제스처 - ## Description 카드를 꾹 눌러서 위/아래로 보내는 제스처 구현 ## ScreenShot ## To-do - [ ] 카드를 꾹 누르는 제스처 + 드래그 제스처 합치기 - [ ] 카드를 위로 보내기 - [ ] 카드를 아래로 내리기 ## Etc
non_main
카드 long press swipe 제스처 description 카드를 꾹 눌러서 위 아래로 보내는 제스처 구현 screenshot to do 카드를 꾹 누르는 제스처 드래그 제스처 합치기 카드를 위로 보내기 카드를 아래로 내리기 etc
0
231,033
25,482,873,244
IssuesEvent
2022-11-26 01:48:20
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2017-17712 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2017-17712 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The raw_sendmsg() function in net/ipv4/raw.c in the Linux kernel through 4.14.6 has a race condition in inet->hdrincl that leads to uninitialized stack pointer usage; this allows a local user to execute code and gain privileges. <p>Publish Date: 2017-12-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-17712>CVE-2017-17712</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-17712">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-17712</a></p> <p>Release Date: 2017-12-16</p> <p>Fix Resolution: v4.15-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-17712 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-17712 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The raw_sendmsg() function in net/ipv4/raw.c in the Linux kernel through 4.14.6 has a race condition in inet->hdrincl that leads to uninitialized stack pointer usage; this allows a local user to execute code and gain privileges. <p>Publish Date: 2017-12-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-17712>CVE-2017-17712</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-17712">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-17712</a></p> <p>Release Date: 2017-12-16</p> <p>Fix Resolution: v4.15-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the raw sendmsg function in net raw c in the linux kernel through has a race condition in inet hdrincl that leads to uninitialized stack pointer usage this allows a local user to execute code and gain privileges publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,286
8,139,921,915
IssuesEvent
2018-08-20 19:19:26
AlexsLemonade/refinebio-frontend
https://api.github.com/repos/AlexsLemonade/refinebio-frontend
closed
Refactor: Use the url as the source of truth for search filters
backlog maintainability
In the search page, we're parsing the filters from the url and copying them into the Redux store. Then when a filter is applied or a new search term is entered, the changed is saved into the Redux store and then pushed into the url. We can simplify the code if we use the url as the single source of truth for the filters and the search term, instead of having to maintain the store and the url synced.
True
Refactor: Use the url as the source of truth for search filters - In the search page, we're parsing the filters from the url and copying them into the Redux store. Then when a filter is applied or a new search term is entered, the changed is saved into the Redux store and then pushed into the url. We can simplify the code if we use the url as the single source of truth for the filters and the search term, instead of having to maintain the store and the url synced.
main
refactor use the url as the source of truth for search filters in the search page we re parsing the filters from the url and copying them into the redux store then when a filter is applied or a new search term is entered the changed is saved into the redux store and then pushed into the url we can simplify the code if we use the url as the single source of truth for the filters and the search term instead of having to maintain the store and the url synced
1
110,449
23,934,169,359
IssuesEvent
2022-09-11 01:19:37
Pokecube-Development/Pokecube-Issues-and-Wiki
https://api.github.com/repos/Pokecube-Development/Pokecube-Issues-and-Wiki
closed
Wearables Render Incorrectly
Bug - Code Fixed
#### Issue Description: - Wearables render incorrectly in the gui - The wearables button does not disappear when switching tabs in creative. #### What happens: - When a wearable is equipped, it appears inverted. - The wearables button does not disappear when switching tabs in creative. #### What you expected to happen: - Wearables should render correctly. #### Steps to reproduce: 1. Equip a wearable 2. Observe in gui ... ____ #### Affected Versions: - Pokecube AIO: dev - Minecraft: 1.18.1 - Forge: 39.0.19 ![2022-01-24_20 24 54](https://user-images.githubusercontent.com/12000593/150909980-9d18b967-1231-4382-b950-b4207907cf3a.png) ![2022-01-24_22 28 22](https://user-images.githubusercontent.com/12000593/150911152-1b08bfc7-4af2-47f8-bff8-4638efdbd937.png)
1.0
Wearables Render Incorrectly - #### Issue Description: - Wearables render incorrectly in the gui - The wearables button does not disappear when switching tabs in creative. #### What happens: - When a wearable is equipped, it appears inverted. - The wearables button does not disappear when switching tabs in creative. #### What you expected to happen: - Wearables should render correctly. #### Steps to reproduce: 1. Equip a wearable 2. Observe in gui ... ____ #### Affected Versions: - Pokecube AIO: dev - Minecraft: 1.18.1 - Forge: 39.0.19 ![2022-01-24_20 24 54](https://user-images.githubusercontent.com/12000593/150909980-9d18b967-1231-4382-b950-b4207907cf3a.png) ![2022-01-24_22 28 22](https://user-images.githubusercontent.com/12000593/150911152-1b08bfc7-4af2-47f8-bff8-4638efdbd937.png)
non_main
wearables render incorrectly issue description wearables render incorrectly in the gui the wearables button does not disappear when switching tabs in creative what happens when a wearable is equipped it appears inverted the wearables button does not disappear when switching tabs in creative what you expected to happen wearables should render correctly steps to reproduce equip a wearable observe in gui affected versions pokecube aio dev minecraft forge
0
135,506
5,253,631,237
IssuesEvent
2017-02-02 10:14:01
NRGI/resourcecontracts.org
https://api.github.com/repos/NRGI/resourcecontracts.org
closed
Contracts not annotated noted as annotated
bug Priority RC country request
Search for all Tunisia contracts annotated shows 54 contracts on both the [public ](http://www.resourcecontracts.org/search?q=&country%5B%5D=tn&annotated=1)and [country ](http://tunisia.resourcecontracts.org/search?q=&annotated=1)site. The issue is that the search is displaying contracts that haven't been annotated like this [contract](http://admin.resourcecontracts.org/contract/2270).
1.0
Contracts not annotated noted as annotated - Search for all Tunisia contracts annotated shows 54 contracts on both the [public ](http://www.resourcecontracts.org/search?q=&country%5B%5D=tn&annotated=1)and [country ](http://tunisia.resourcecontracts.org/search?q=&annotated=1)site. The issue is that the search is displaying contracts that haven't been annotated like this [contract](http://admin.resourcecontracts.org/contract/2270).
non_main
contracts not annotated noted as annotated search for all tunisia contracts annotated shows contracts on both the the issue is that the search is displaying contracts that haven t been annotated like this
0
591,579
17,851,842,255
IssuesEvent
2021-09-04 07:57:50
ballerina-platform/ballerina-dev-website
https://api.github.com/repos/ballerina-platform/ballerina-dev-website
closed
For ballerina central instad of contribute, we need to use publish
Priority/Highest Area/Docs Type/Bug Points/0.5
**Description:** $subject in [1]. It should be `Publish to Ballerina Central`. Also saying module is not correct and we need to change it to package. [1] https://dev.ballerina.io/community/ <img width="1284" alt="image" src="https://user-images.githubusercontent.com/16300038/130505889-48cda70c-8b4e-4631-8240-f0331a9d66f1.png"> **Steps to reproduce:** **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
For ballerina central instad of contribute, we need to use publish - **Description:** $subject in [1]. It should be `Publish to Ballerina Central`. Also saying module is not correct and we need to change it to package. [1] https://dev.ballerina.io/community/ <img width="1284" alt="image" src="https://user-images.githubusercontent.com/16300038/130505889-48cda70c-8b4e-4631-8240-f0331a9d66f1.png"> **Steps to reproduce:** **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
non_main
for ballerina central instad of contribute we need to use publish description subject in it should be publish to ballerina central also saying module is not correct and we need to change it to package img width alt image src steps to reproduce affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
0
144,422
19,286,246,692
IssuesEvent
2021-12-11 02:09:02
YauheniPo/PDF-Parser-HTTPS-client
https://api.github.com/repos/YauheniPo/PDF-Parser-HTTPS-client
opened
CVE-2021-44228 (High) detected in log4j-core-2.13.2.jar
security vulnerability
## CVE-2021-44228 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.13.2.jar</b></p></summary> <p>The Apache Log4j Implementation</p> <p>Library home page: <a href="https://logging.apache.org/log4j/2.x/">https://logging.apache.org/log4j/2.x/</a></p> <p>Path to dependency file: PDF-Parser-HTTPS-client/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/apache/logging/log4j/log4j-core/2.13.2/log4j-core-2.13.2.jar</p> <p> Dependency Hierarchy: - :x: **log4j-core-2.13.2.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Log4j versions prior to 2.15.0 are subject to a remote code execution vulnerability via the ldap JNDI parser. <p>Publish Date: 2021-11-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44228>CVE-2021-44228</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>10.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jfh8-c2jp-5v3q">https://github.com/advisories/GHSA-jfh8-c2jp-5v3q</a></p> <p>Release Date: 2021-12-10</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.15.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-44228 (High) detected in log4j-core-2.13.2.jar - ## CVE-2021-44228 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.13.2.jar</b></p></summary> <p>The Apache Log4j Implementation</p> <p>Library home page: <a href="https://logging.apache.org/log4j/2.x/">https://logging.apache.org/log4j/2.x/</a></p> <p>Path to dependency file: PDF-Parser-HTTPS-client/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/apache/logging/log4j/log4j-core/2.13.2/log4j-core-2.13.2.jar</p> <p> Dependency Hierarchy: - :x: **log4j-core-2.13.2.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Log4j versions prior to 2.15.0 are subject to a remote code execution vulnerability via the ldap JNDI parser. <p>Publish Date: 2021-11-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44228>CVE-2021-44228</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>10.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-jfh8-c2jp-5v3q">https://github.com/advisories/GHSA-jfh8-c2jp-5v3q</a></p> <p>Release Date: 2021-12-10</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.15.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in core jar cve high severity vulnerability vulnerable library core jar the apache implementation library home page a href path to dependency file pdf parser https client pom xml path to vulnerable library canner repository org apache logging core core jar dependency hierarchy x core jar vulnerable library found in base branch master vulnerability details versions prior to are subject to a remote code execution vulnerability via the ldap jndi parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core step up your open source security game with whitesource
0
4,791
24,675,175,669
IssuesEvent
2022-10-18 16:24:30
bazelbuild/intellij
https://api.github.com/repos/bazelbuild/intellij
closed
External libraries do not maintain package structure
type: bug lang: go product: GoLand topic: external dependencies awaiting-maintainer
![Screen Shot 2021-02-12 at 12 28 33 PM](https://user-images.githubusercontent.com/7622243/107820329-7a704280-6d2f-11eb-8b4f-423eb61507e4.png) When using just go modules (not bazel) and you click on the “source” for a given import, you’re brought to the import package. With bazel, all the deps are single files — we lose the package structure of the deps. Viewing the deps in a given package is a useful feature. Is there a way to retain this functionality using bazel+goland?
True
External libraries do not maintain package structure - ![Screen Shot 2021-02-12 at 12 28 33 PM](https://user-images.githubusercontent.com/7622243/107820329-7a704280-6d2f-11eb-8b4f-423eb61507e4.png) When using just go modules (not bazel) and you click on the “source” for a given import, you’re brought to the import package. With bazel, all the deps are single files — we lose the package structure of the deps. Viewing the deps in a given package is a useful feature. Is there a way to retain this functionality using bazel+goland?
main
external libraries do not maintain package structure when using just go modules not bazel and you click on the “source” for a given import you’re brought to the import package with bazel all the deps are single files — we lose the package structure of the deps viewing the deps in a given package is a useful feature is there a way to retain this functionality using bazel goland
1
3,930
17,690,683,017
IssuesEvent
2021-08-24 09:33:10
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
Hovering over the selected item from a dropdown doesn't show the complete text
type: bug 🐛 status: waiting for maintainer response 💬
## What package(s) are you using? - [x] `carbon-components-react` ## Description > Is this issue related to a specific component? `Dropdown`, `Combobox` > What did you expect to happen? What happened instead? What would you like to > see changed? It is expected to see the entire text when we hover over a selected option from the dropdown. Like it shows the full text when we hover over any option when the dropdown is open. > What browser are you working in? Chrome ## Steps to reproduce the issue For Dropdown: 1. Go to [Carbon Components React](https://react.carbondesignsystem.com/?path=/story/components-dropdown--default) portal, `Dropdown` page 2. Select the second option from the dropdown list - `Lorem, ipsum dolor sit amet consectetur adipisicing elit.` 3. Because of its length the string is truncated and it is expected to show the entire string when we hover over it. But it is not the case in the current implementation. For Combobox: 1. Go to [Carbon Components React](https://react.carbondesignsystem.com/?path=/story/components-combobox--combobox) portal, `Combobox` page 2. Select the second option from the dropdown list - `An example option that is really long to show what should be done to handle long text` 3. Because of its length the string is truncated and it is expected to show the entire string when we hover over it. But it is not the case in the current implementation.
True
Hovering over the selected item from a dropdown doesn't show the complete text - ## What package(s) are you using? - [x] `carbon-components-react` ## Description > Is this issue related to a specific component? `Dropdown`, `Combobox` > What did you expect to happen? What happened instead? What would you like to > see changed? It is expected to see the entire text when we hover over a selected option from the dropdown. Like it shows the full text when we hover over any option when the dropdown is open. > What browser are you working in? Chrome ## Steps to reproduce the issue For Dropdown: 1. Go to [Carbon Components React](https://react.carbondesignsystem.com/?path=/story/components-dropdown--default) portal, `Dropdown` page 2. Select the second option from the dropdown list - `Lorem, ipsum dolor sit amet consectetur adipisicing elit.` 3. Because of its length the string is truncated and it is expected to show the entire string when we hover over it. But it is not the case in the current implementation. For Combobox: 1. Go to [Carbon Components React](https://react.carbondesignsystem.com/?path=/story/components-combobox--combobox) portal, `Combobox` page 2. Select the second option from the dropdown list - `An example option that is really long to show what should be done to handle long text` 3. Because of its length the string is truncated and it is expected to show the entire string when we hover over it. But it is not the case in the current implementation.
main
hovering over the selected item from a dropdown doesn t show the complete text what package s are you using carbon components react description is this issue related to a specific component dropdown combobox what did you expect to happen what happened instead what would you like to see changed it is expected to see the entire text when we hover over a selected option from the dropdown like it shows the full text when we hover over any option when the dropdown is open what browser are you working in chrome steps to reproduce the issue for dropdown go to portal dropdown page select the second option from the dropdown list lorem ipsum dolor sit amet consectetur adipisicing elit because of its length the string is truncated and it is expected to show the entire string when we hover over it but it is not the case in the current implementation for combobox go to portal combobox page select the second option from the dropdown list an example option that is really long to show what should be done to handle long text because of its length the string is truncated and it is expected to show the entire string when we hover over it but it is not the case in the current implementation
1
810
4,434,218,600
IssuesEvent
2016-08-18 01:13:02
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
get_url not using environment variable no_proxy
bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME get_url ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### CONFIGURATION default ##### OS / ENVIRONMENT Red Hat Enterprise Linux Server release 7.2 (Maipo) ##### SUMMARY When setting the environment variable no_proxy, the get_url module doesn't use it. We need to bypass our corporate proxy to get files from our central fileshare. ##### STEPS TO REPRODUCE ``` --- - hosts: localhost vars: proxy_url: "http://aws-proxy-us-east-1.company.com:8080" noproxy_url: "127.0.0.1,localhost,.local,169.254.169.254,.fileshare.company.com" url_to_get: "https://fileshare.company.com/filename.tar" dest_dir: "/tmp" tasks: - debug: msg="proxy_url is {{ proxy_url }}" - debug: msg="noproxy_url is {{ noproxy_url }}" - debug: msg="url_to_get is {{ url_to_get }}" - name: "Set proxy environment variables" set_fact: environment_vars: http_proxy: "{{ proxy_url }}" https_proxy: "{{ proxy_url }}" no_proxy: "{{ noproxy_url }}" HTTP_PROXY: "{{ proxy_url }}" HTTPS_PROXY: "{{ proxy_url }}" NO_PROXY: "{{ noproxy_url }}" - name: "Download something with no env vars set" local_action: module: get_url url: "{{ url_to_get }}" dest: "{{ dest_dir }}/wtf" validate_certs: no force: yes become: no - name: "Download something with env vars set" local_action: module: get_url url: "{{ url_to_get }}" dest: "{{ dest_dir }}/wtf2" validate_certs: no force: yes environment: "{{ environment_vars }}" become: no register: get_url_err ignore_errors: yes - debug: var=get_url_err ``` ##### EXPECTED RESULTS Both should download the file and drop in /tmp. For comparison, I ran this from command line using "curl". When I set the http_proxy and https_proxy, it failed with a 404 (blocked by the proxy). When I set no_proxy to our fileshare, it worked (bypassed the proxy). ##### ACTUAL RESULTS First task worked, second one says it timed out. However, when I captured the module output to a variable and dumped it, it was actually a 404. ``` TASK [Download something with env vars set] ************************************ task path: /home/ec2-user/geturltest.yml:34 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: ec2-user <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946 `" && echo ansible-tmp-1465584747.19-272127080396946="` echo $HOME/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946 `" ) && sleep 0' <localhost> PUT /tmp/tmpC_cHj6 TO /home/ec2-user/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946/get_url <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 HTTP_PROXY=http://aws-proxy-us-east-1.company.com:8080 LC_MESSAGES=en_US.UTF-8 HTTPS_PROXY=http://aws-proxy-us-east-1.company.com:8080 NO_PROXY=127.0.0.1,localhost,.local,169.254.169.254,.fileshare.company.com http_proxy=http://aws-proxy-us-east-1.company.com:8080 https_proxy=http://aws-proxy-us-east-1.company.com:8080 LC_ALL=en_US.UTF-8 no_proxy=127.0.0.1,localhost,.local,169.254.169.254,.fileshare.company.com /usr/bin/python /home/ec2-user/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946/get_url; rm -rf "/home/ec2-user/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946/" > /dev/null 2>&1 && sleep 0' fatal: [localhost -> localhost]: FAILED! => {"changed": false, "dest": "/tmp/wtf2", "failed": true, "gid": 1000, "group": "ec2-user", "invocation": {"module_args": {"backup": false, "checksum": "", "content": null, "delimiter": null, "dest": "/tmp/wtf2", "directory_mode": null, "follow": false, "force": true, "force_basic_auth": false, "group": null, "headers": null, "http_agent": "ansible-httpget", "mode": null, "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "sha256sum": "", "src": null, "timeout": 10, "tmp_dest": "", "url": "https://fileshare.company.com/filename.tar", "url_password": null, "url_username": null, "use_proxy": true, "validate_certs": false}, "module_name": "get_url"}, "mode": "0644", "msg": "Request failed", "owner": "ec2-user", "response": "Request failed: <urlopen error timed out>", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 74485760, "state": "file", "status_code": -1, "uid": 1000, "url": "https://fileshare.company.com/filename.tar"} TASK [debug] ******************************************************************* task path: /home/ec2-user/geturltest.yml:47 ok: [localhost] => { "get_url_err": { "changed": false, "dest": "/tmp/wtf2", "failed": true, "gid": 1000, "group": "ec2-user", "mode": "0644", "msg": "Request failed", "owner": "ec2-user", "response": "HTTP Error 403: Forbidden", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 74485760, "state": "file", "status_code": 403, "uid": 1000, "url": "https://fileshare.company.com/filename.tar" } } ```
True
get_url not using environment variable no_proxy - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME get_url ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### CONFIGURATION default ##### OS / ENVIRONMENT Red Hat Enterprise Linux Server release 7.2 (Maipo) ##### SUMMARY When setting the environment variable no_proxy, the get_url module doesn't use it. We need to bypass our corporate proxy to get files from our central fileshare. ##### STEPS TO REPRODUCE ``` --- - hosts: localhost vars: proxy_url: "http://aws-proxy-us-east-1.company.com:8080" noproxy_url: "127.0.0.1,localhost,.local,169.254.169.254,.fileshare.company.com" url_to_get: "https://fileshare.company.com/filename.tar" dest_dir: "/tmp" tasks: - debug: msg="proxy_url is {{ proxy_url }}" - debug: msg="noproxy_url is {{ noproxy_url }}" - debug: msg="url_to_get is {{ url_to_get }}" - name: "Set proxy environment variables" set_fact: environment_vars: http_proxy: "{{ proxy_url }}" https_proxy: "{{ proxy_url }}" no_proxy: "{{ noproxy_url }}" HTTP_PROXY: "{{ proxy_url }}" HTTPS_PROXY: "{{ proxy_url }}" NO_PROXY: "{{ noproxy_url }}" - name: "Download something with no env vars set" local_action: module: get_url url: "{{ url_to_get }}" dest: "{{ dest_dir }}/wtf" validate_certs: no force: yes become: no - name: "Download something with env vars set" local_action: module: get_url url: "{{ url_to_get }}" dest: "{{ dest_dir }}/wtf2" validate_certs: no force: yes environment: "{{ environment_vars }}" become: no register: get_url_err ignore_errors: yes - debug: var=get_url_err ``` ##### EXPECTED RESULTS Both should download the file and drop in /tmp. For comparison, I ran this from command line using "curl". When I set the http_proxy and https_proxy, it failed with a 404 (blocked by the proxy). When I set no_proxy to our fileshare, it worked (bypassed the proxy). ##### ACTUAL RESULTS First task worked, second one says it timed out. However, when I captured the module output to a variable and dumped it, it was actually a 404. ``` TASK [Download something with env vars set] ************************************ task path: /home/ec2-user/geturltest.yml:34 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: ec2-user <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946 `" && echo ansible-tmp-1465584747.19-272127080396946="` echo $HOME/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946 `" ) && sleep 0' <localhost> PUT /tmp/tmpC_cHj6 TO /home/ec2-user/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946/get_url <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 HTTP_PROXY=http://aws-proxy-us-east-1.company.com:8080 LC_MESSAGES=en_US.UTF-8 HTTPS_PROXY=http://aws-proxy-us-east-1.company.com:8080 NO_PROXY=127.0.0.1,localhost,.local,169.254.169.254,.fileshare.company.com http_proxy=http://aws-proxy-us-east-1.company.com:8080 https_proxy=http://aws-proxy-us-east-1.company.com:8080 LC_ALL=en_US.UTF-8 no_proxy=127.0.0.1,localhost,.local,169.254.169.254,.fileshare.company.com /usr/bin/python /home/ec2-user/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946/get_url; rm -rf "/home/ec2-user/.ansible/tmp/ansible-tmp-1465584747.19-272127080396946/" > /dev/null 2>&1 && sleep 0' fatal: [localhost -> localhost]: FAILED! => {"changed": false, "dest": "/tmp/wtf2", "failed": true, "gid": 1000, "group": "ec2-user", "invocation": {"module_args": {"backup": false, "checksum": "", "content": null, "delimiter": null, "dest": "/tmp/wtf2", "directory_mode": null, "follow": false, "force": true, "force_basic_auth": false, "group": null, "headers": null, "http_agent": "ansible-httpget", "mode": null, "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "sha256sum": "", "src": null, "timeout": 10, "tmp_dest": "", "url": "https://fileshare.company.com/filename.tar", "url_password": null, "url_username": null, "use_proxy": true, "validate_certs": false}, "module_name": "get_url"}, "mode": "0644", "msg": "Request failed", "owner": "ec2-user", "response": "Request failed: <urlopen error timed out>", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 74485760, "state": "file", "status_code": -1, "uid": 1000, "url": "https://fileshare.company.com/filename.tar"} TASK [debug] ******************************************************************* task path: /home/ec2-user/geturltest.yml:47 ok: [localhost] => { "get_url_err": { "changed": false, "dest": "/tmp/wtf2", "failed": true, "gid": 1000, "group": "ec2-user", "mode": "0644", "msg": "Request failed", "owner": "ec2-user", "response": "HTTP Error 403: Forbidden", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 74485760, "state": "file", "status_code": 403, "uid": 1000, "url": "https://fileshare.company.com/filename.tar" } } ```
main
get url not using environment variable no proxy issue type bug report component name get url ansible version ansible configuration default os environment red hat enterprise linux server release maipo summary when setting the environment variable no proxy the get url module doesn t use it we need to bypass our corporate proxy to get files from our central fileshare steps to reproduce hosts localhost vars proxy url noproxy url localhost local fileshare company com url to get dest dir tmp tasks debug msg proxy url is proxy url debug msg noproxy url is noproxy url debug msg url to get is url to get name set proxy environment variables set fact environment vars http proxy proxy url https proxy proxy url no proxy noproxy url http proxy proxy url https proxy proxy url no proxy noproxy url name download something with no env vars set local action module get url url url to get dest dest dir wtf validate certs no force yes become no name download something with env vars set local action module get url url url to get dest dest dir validate certs no force yes environment environment vars become no register get url err ignore errors yes debug var get url err expected results both should download the file and drop in tmp for comparison i ran this from command line using curl when i set the http proxy and https proxy it failed with a blocked by the proxy when i set no proxy to our fileshare it worked bypassed the proxy actual results first task worked second one says it timed out however when i captured the module output to a variable and dumped it it was actually a task task path home user geturltest yml establish local connection for user user exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpc to home user ansible tmp ansible tmp get url exec bin sh c lang en us utf http proxy lc messages en us utf https proxy no proxy localhost local fileshare company com http proxy https proxy lc all en us utf no proxy localhost local fileshare company com usr bin python home user ansible tmp ansible tmp get url rm rf home user ansible tmp ansible tmp dev null sleep fatal failed changed false dest tmp failed true gid group user invocation module args backup false checksum content null delimiter null dest tmp directory mode null follow false force true force basic auth false group null headers null http agent ansible httpget mode null owner null regexp null remote src null selevel null serole null setype null seuser null src null timeout tmp dest url url password null url username null use proxy true validate certs false module name get url mode msg request failed owner user response request failed secontext unconfined u object r user tmp t size state file status code uid url task task path home user geturltest yml ok get url err changed false dest tmp failed true gid group user mode msg request failed owner user response http error forbidden secontext unconfined u object r user tmp t size state file status code uid url
1
379,313
11,219,841,016
IssuesEvent
2020-01-07 14:42:04
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
Remove Knative Build and JFR from Jenkins X installation
area/fox area/install area/knative kind/enhancement priority/important-longterm
### Summary The knative build and JFR should be removed from Jenkins X installation, as soon as can be used as a buildpack in JenkinsX/Tekton pipeline. #4229 The the moment this topology is using a old fork of prow which is unmaintained. ### Steps to reproduce the behavior ### Expected behavior ### Actual behavior ### Jx version The output of `jx version` is: ``` COPY OUTPUT HERE ``` ### Jenkins type <!-- Select which installation type are you using. --> - [ ] Next Generation (Tekton + Prow) - [ ] Classic Jenkins - [x] Serverless Jenkins (JenkinsFileRunner + Prow) ### Kubernetes cluster <!-- What kind of Kubernetes cluster are you using & how did you create it? --> ### Operating system / Environment <!-- In which environment are you running the jx CLI? -->
1.0
Remove Knative Build and JFR from Jenkins X installation - ### Summary The knative build and JFR should be removed from Jenkins X installation, as soon as can be used as a buildpack in JenkinsX/Tekton pipeline. #4229 The the moment this topology is using a old fork of prow which is unmaintained. ### Steps to reproduce the behavior ### Expected behavior ### Actual behavior ### Jx version The output of `jx version` is: ``` COPY OUTPUT HERE ``` ### Jenkins type <!-- Select which installation type are you using. --> - [ ] Next Generation (Tekton + Prow) - [ ] Classic Jenkins - [x] Serverless Jenkins (JenkinsFileRunner + Prow) ### Kubernetes cluster <!-- What kind of Kubernetes cluster are you using & how did you create it? --> ### Operating system / Environment <!-- In which environment are you running the jx CLI? -->
non_main
remove knative build and jfr from jenkins x installation summary the knative build and jfr should be removed from jenkins x installation as soon as can be used as a buildpack in jenkinsx tekton pipeline the the moment this topology is using a old fork of prow which is unmaintained steps to reproduce the behavior expected behavior actual behavior jx version the output of jx version is copy output here jenkins type select which installation type are you using next generation tekton prow classic jenkins serverless jenkins jenkinsfilerunner prow kubernetes cluster what kind of kubernetes cluster are you using how did you create it operating system environment in which environment are you running the jx cli
0
50,581
12,530,053,274
IssuesEvent
2020-06-04 12:26:51
pravega/pravega
https://api.github.com/repos/pravega/pravega
closed
Pravega 0.8 does not build on JDK14 - Tools issues
area/build version/0.8.0
**Problem description** You cannot build current master of Pravega with JDK14, due to a compatibility issue with Spotbugs,Mockito and JaCoCo **Problem location** Current versions of Spotbugs,Mockito and JaCoCo do not work with JDK14 **Suggestions for an improvement** Upgrade to latest versions of Spotbugs,Mockito and JaCoCo
1.0
Pravega 0.8 does not build on JDK14 - Tools issues - **Problem description** You cannot build current master of Pravega with JDK14, due to a compatibility issue with Spotbugs,Mockito and JaCoCo **Problem location** Current versions of Spotbugs,Mockito and JaCoCo do not work with JDK14 **Suggestions for an improvement** Upgrade to latest versions of Spotbugs,Mockito and JaCoCo
non_main
pravega does not build on tools issues problem description you cannot build current master of pravega with due to a compatibility issue with spotbugs mockito and jacoco problem location current versions of spotbugs mockito and jacoco do not work with suggestions for an improvement upgrade to latest versions of spotbugs mockito and jacoco
0
666
4,195,189,345
IssuesEvent
2016-06-25 15:29:41
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
New Duration Calculator Goodie
Maintainer Approved Suggestion
A simple calculator that enables users to enter queries of time like 3 hours 45 minutes + 1 hour 20 minutes and returns the calculated result. --- IA Page: https://duck.co/ia/view/duration_calculator
True
New Duration Calculator Goodie - A simple calculator that enables users to enter queries of time like 3 hours 45 minutes + 1 hour 20 minutes and returns the calculated result. --- IA Page: https://duck.co/ia/view/duration_calculator
main
new duration calculator goodie a simple calculator that enables users to enter queries of time like hours minutes hour minutes and returns the calculated result ia page
1
134,519
19,252,745,226
IssuesEvent
2021-12-09 07:57:56
QUT-Motorsport/QUTMS_Driverless
https://api.github.com/repos/QUT-Motorsport/QUTMS_Driverless
closed
Intermediate control algorithm
Design 2021
Build upon basic FSDS control algorithm (from industry night) to better represent launch demo. Might have to change lots due to real cone data processing
1.0
Intermediate control algorithm - Build upon basic FSDS control algorithm (from industry night) to better represent launch demo. Might have to change lots due to real cone data processing
non_main
intermediate control algorithm build upon basic fsds control algorithm from industry night to better represent launch demo might have to change lots due to real cone data processing
0
263
3,020,963,656
IssuesEvent
2015-07-31 11:47:40
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Uninstallation of graphviz fails
bug cask waiting-maintainer-feedback
``` brew cask uninstall graphviz ==> Running uninstall process for graphviz; your password may be necessary ==> Removing files from pkgutil Bill-of-Materials ==> rmdir: /usr/local/include/graphviz: Not a directory Error: Command failed to execute! ==> Failed command: ["/usr/bin/sudo", "-E", "--", "/bin/rmdir", "--", "#<Pathname:/usr/local/include/graphviz>"] ==> Output of failed command: ==> Exit status of failed command: #<Process::Status: pid 70281 exit 1> ```
True
Uninstallation of graphviz fails - ``` brew cask uninstall graphviz ==> Running uninstall process for graphviz; your password may be necessary ==> Removing files from pkgutil Bill-of-Materials ==> rmdir: /usr/local/include/graphviz: Not a directory Error: Command failed to execute! ==> Failed command: ["/usr/bin/sudo", "-E", "--", "/bin/rmdir", "--", "#<Pathname:/usr/local/include/graphviz>"] ==> Output of failed command: ==> Exit status of failed command: #<Process::Status: pid 70281 exit 1> ```
main
uninstallation of graphviz fails brew cask uninstall graphviz running uninstall process for graphviz your password may be necessary removing files from pkgutil bill of materials rmdir usr local include graphviz not a directory error command failed to execute failed command output of failed command exit status of failed command
1
166,639
20,719,124,149
IssuesEvent
2022-03-13 04:46:22
TIBCOSoftware/justapis-javascript-sdk
https://api.github.com/repos/TIBCOSoftware/justapis-javascript-sdk
closed
WS-2014-0005 (High) detected in qs-0.5.6.tgz - autoclosed
security vulnerability
## WS-2014-0005 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.5.6.tgz</b></p></summary> <p>querystring parser</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.5.6.tgz">https://registry.npmjs.org/qs/-/qs-0.5.6.tgz</a></p> <p>Path to dependency file: /justapis-javascript-sdk/package.json</p> <p>Path to vulnerable library: /node_modules/qs/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-0.6.1.tgz (Root Library) - tiny-lr-fork-0.0.5.tgz - :x: **qs-0.5.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TIBCOSoftware/justapis-javascript-sdk/commits/3ca192403e92db3173fd513bbb67c49050b748e7">3ca192403e92db3173fd513bbb67c49050b748e7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Denial-of-Service Extended Event Loop Blocking.The qs module does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time <p>Publish Date: 2014-07-31 <p>URL: <a href=https://github.com/ljharb/qs/commit/6667340dd3c7deaa0eb1c27f175faaaf71f19823>WS-2014-0005</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2014-0005">https://nvd.nist.gov/vuln/detail/WS-2014-0005</a></p> <p>Release Date: 2014-07-31</p> <p>Fix Resolution: qs - 1.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.5.6","packageFilePaths":["/justapis-javascript-sdk/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-watch:0.6.1;tiny-lr-fork:0.0.5;qs:0.5.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 1.0.0","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"WS-2014-0005","vulnerabilityDetails":"Denial-of-Service Extended Event Loop Blocking.The qs module does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time","vulnerabilityUrl":"https://github.com/ljharb/qs/commit/6667340dd3c7deaa0eb1c27f175faaaf71f19823","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2014-0005 (High) detected in qs-0.5.6.tgz - autoclosed - ## WS-2014-0005 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-0.5.6.tgz</b></p></summary> <p>querystring parser</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.5.6.tgz">https://registry.npmjs.org/qs/-/qs-0.5.6.tgz</a></p> <p>Path to dependency file: /justapis-javascript-sdk/package.json</p> <p>Path to vulnerable library: /node_modules/qs/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-0.6.1.tgz (Root Library) - tiny-lr-fork-0.0.5.tgz - :x: **qs-0.5.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TIBCOSoftware/justapis-javascript-sdk/commits/3ca192403e92db3173fd513bbb67c49050b748e7">3ca192403e92db3173fd513bbb67c49050b748e7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Denial-of-Service Extended Event Loop Blocking.The qs module does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time <p>Publish Date: 2014-07-31 <p>URL: <a href=https://github.com/ljharb/qs/commit/6667340dd3c7deaa0eb1c27f175faaaf71f19823>WS-2014-0005</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2014-0005">https://nvd.nist.gov/vuln/detail/WS-2014-0005</a></p> <p>Release Date: 2014-07-31</p> <p>Fix Resolution: qs - 1.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.5.6","packageFilePaths":["/justapis-javascript-sdk/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-watch:0.6.1;tiny-lr-fork:0.0.5;qs:0.5.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 1.0.0","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"WS-2014-0005","vulnerabilityDetails":"Denial-of-Service Extended Event Loop Blocking.The qs module does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time","vulnerabilityUrl":"https://github.com/ljharb/qs/commit/6667340dd3c7deaa0eb1c27f175faaaf71f19823","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
ws high detected in qs tgz autoclosed ws high severity vulnerability vulnerable library qs tgz querystring parser library home page a href path to dependency file justapis javascript sdk package json path to vulnerable library node modules qs package json dependency hierarchy grunt contrib watch tgz root library tiny lr fork tgz x qs tgz vulnerable library found in head commit a href vulnerability details denial of service extended event loop blocking the qs module does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt contrib watch tiny lr fork qs isminimumfixversionavailable true minimumfixversion qs isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails denial of service extended event loop blocking the qs module does not have an option or default for specifying object depth and when parsing a string representing a deeply nested object will block the event loop for long periods of time vulnerabilityurl
0
3,651
14,916,358,697
IssuesEvent
2021-01-22 18:04:58
melisMirza/SWE573_project
https://api.github.com/repos/melisMirza/SWE573_project
closed
Schedule Data Collection Job on Heroku
deployment maintainance
Schedule a job for data stream. Same job that is currently running on local can be used. 300 posts every 30 mins.
True
Schedule Data Collection Job on Heroku - Schedule a job for data stream. Same job that is currently running on local can be used. 300 posts every 30 mins.
main
schedule data collection job on heroku schedule a job for data stream same job that is currently running on local can be used posts every mins
1
1,480
6,415,813,901
IssuesEvent
2017-08-08 13:37:30
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
vmware_host is not idempotent
affects_2.3 bug_report cloud vmware waiting_on_maintainer
When adding an host that already exists, it fails with the error: 'msg: 'vim.fault.DuplicateName' object has no attribute 'thumbprint'
True
vmware_host is not idempotent - When adding an host that already exists, it fails with the error: 'msg: 'vim.fault.DuplicateName' object has no attribute 'thumbprint'
main
vmware host is not idempotent when adding an host that already exists it fails with the error msg vim fault duplicatename object has no attribute thumbprint
1
1,312
5,558,697,078
IssuesEvent
2017-03-24 15:18:09
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_service doesn't pass environment to docker-compose
affects_2.1 bug_report cloud docker waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME docker_service ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.2.0 config file = /usr/local/etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> N/A ##### SUMMARY `docker_service` task doesn't pass environment variables defined using `environment` option at task- or playbook-level to the `docker-compose.yml` when using `project_src`. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Playbook: ``` yaml --- - name: demo tasks: - docker_service: project_src=/tmp/compose-dir environment: MYSQL_DB: "test" ``` `docker-compose.yml` at `/tmp/compose-dir`: ``` yaml version: '2' services: mysql: image: mysql:5.7 environment: MYSQL_DATABASE: ${MYSQL_DB} ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS `mysql` container should see `MYSQL_DATABASE=test` and create database "test". ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> "test" database doesn't created But when calling docker-compose directly: ``` MYSQL_DB=test docker-compose up ``` the "test" database is created
True
docker_service doesn't pass environment to docker-compose - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME docker_service ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.2.0 config file = /usr/local/etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> N/A ##### SUMMARY `docker_service` task doesn't pass environment variables defined using `environment` option at task- or playbook-level to the `docker-compose.yml` when using `project_src`. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Playbook: ``` yaml --- - name: demo tasks: - docker_service: project_src=/tmp/compose-dir environment: MYSQL_DB: "test" ``` `docker-compose.yml` at `/tmp/compose-dir`: ``` yaml version: '2' services: mysql: image: mysql:5.7 environment: MYSQL_DATABASE: ${MYSQL_DB} ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS `mysql` container should see `MYSQL_DATABASE=test` and create database "test". ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> "test" database doesn't created But when calling docker-compose directly: ``` MYSQL_DB=test docker-compose up ``` the "test" database is created
main
docker service doesn t pass environment to docker compose issue type bug report component name docker service ansible version ansible config file usr local etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific n a summary docker service task doesn t pass environment variables defined using environment option at task or playbook level to the docker compose yml when using project src steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used playbook yaml name demo tasks docker service project src tmp compose dir environment mysql db test docker compose yml at tmp compose dir yaml version services mysql image mysql environment mysql database mysql db expected results mysql container should see mysql database test and create database test actual results test database doesn t created but when calling docker compose directly mysql db test docker compose up the test database is created
1
5,845
31,028,302,932
IssuesEvent
2023-08-10 10:39:00
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
opened
Maintainer change request: Parsedown filter
Maintainer change request
**Thank you for supporting the Backdrop community!** Please note the procedure to add a new maintainer to a project: 1. Please join the Backdrop Contrib group (if you have not already) by submitting [an application](https://github.com/backdrop-ops/contrib/issues/new?assignees=klonos&labels=Maintainer+application&template=application-to-join-the-contrib-group.md&title=Application+to+join+the+Contrib+Group%3A). 2. File an issue in the current project's issue queue offering to help maintain that project. 3. Create a PR for that project that adds your name to the README.md file in the list of maintainers. <!-- The project maintainer, or a backdrop-contrib administrator, will merge this PR to accept your offer of help. --> 4. If the project does not have a listed maintainer, or if a current maintainer does not respond within 2 weeks, create *this issue* to take over the project. **Please include a link to the issue you filed for the project.** https://github.com/backdrop-contrib/parsedown_filter/issues/5 **Please include a link to the PR that adds your name to the README.md file.** https://github.com/backdrop-contrib/parsedown_filter/pull/6 <!-- After confirming the project has been abandoned for a period of 2 weeks or more, a Backdrop Contrib administrator will add your name to the list of maintainers in that project's README.md file, and grant you admin access to the project. -->
True
Maintainer change request: Parsedown filter - **Thank you for supporting the Backdrop community!** Please note the procedure to add a new maintainer to a project: 1. Please join the Backdrop Contrib group (if you have not already) by submitting [an application](https://github.com/backdrop-ops/contrib/issues/new?assignees=klonos&labels=Maintainer+application&template=application-to-join-the-contrib-group.md&title=Application+to+join+the+Contrib+Group%3A). 2. File an issue in the current project's issue queue offering to help maintain that project. 3. Create a PR for that project that adds your name to the README.md file in the list of maintainers. <!-- The project maintainer, or a backdrop-contrib administrator, will merge this PR to accept your offer of help. --> 4. If the project does not have a listed maintainer, or if a current maintainer does not respond within 2 weeks, create *this issue* to take over the project. **Please include a link to the issue you filed for the project.** https://github.com/backdrop-contrib/parsedown_filter/issues/5 **Please include a link to the PR that adds your name to the README.md file.** https://github.com/backdrop-contrib/parsedown_filter/pull/6 <!-- After confirming the project has been abandoned for a period of 2 weeks or more, a Backdrop Contrib administrator will add your name to the list of maintainers in that project's README.md file, and grant you admin access to the project. -->
main
maintainer change request parsedown filter thank you for supporting the backdrop community please note the procedure to add a new maintainer to a project please join the backdrop contrib group if you have not already by submitting file an issue in the current project s issue queue offering to help maintain that project create a pr for that project that adds your name to the readme md file in the list of maintainers the project maintainer or a backdrop contrib administrator will merge this pr to accept your offer of help if the project does not have a listed maintainer or if a current maintainer does not respond within weeks create this issue to take over the project please include a link to the issue you filed for the project please include a link to the pr that adds your name to the readme md file after confirming the project has been abandoned for a period of weeks or more a backdrop contrib administrator will add your name to the list of maintainers in that project s readme md file and grant you admin access to the project
1
220,481
7,360,267,253
IssuesEvent
2018-03-10 16:51:36
systers/portal
https://api.github.com/repos/systers/portal
closed
Password change
Difficulty: MEDIUM Priority: LOW Program: Outreachy Type: Enhancement
Should display the message "The new password cannot be same as the old one" when new password and older password is same.
1.0
Password change - Should display the message "The new password cannot be same as the old one" when new password and older password is same.
non_main
password change should display the message the new password cannot be same as the old one when new password and older password is same
0
211,574
23,833,154,671
IssuesEvent
2022-09-06 01:08:50
joshbnewton31080/WebGoat
https://api.github.com/repos/joshbnewton31080/WebGoat
opened
CVE-2022-38749 (Medium) detected in snakeyaml-1.28.jar
security vulnerability
## CVE-2022-38749 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.28.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /webwolf/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-validation-2.5.4.jar (Root Library) - spring-boot-starter-2.5.4.jar - :x: **snakeyaml-1.28.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/joshbnewton31080/WebGoat/commit/1286cb94ee13fa17303884471f091b9cb58d3c12">1286cb94ee13fa17303884471f091b9cb58d3c12</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38749>CVE-2022-38749</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open">https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p>
True
CVE-2022-38749 (Medium) detected in snakeyaml-1.28.jar - ## CVE-2022-38749 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.28.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /webwolf/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-validation-2.5.4.jar (Root Library) - spring-boot-starter-2.5.4.jar - :x: **snakeyaml-1.28.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/joshbnewton31080/WebGoat/commit/1286cb94ee13fa17303884471f091b9cb58d3c12">1286cb94ee13fa17303884471f091b9cb58d3c12</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38749>CVE-2022-38749</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open">https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p>
non_main
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file webwolf pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter validation jar root library spring boot starter jar x snakeyaml jar vulnerable library found in head commit a href found in base branch develop vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml
0
805,341
29,517,912,685
IssuesEvent
2023-06-04 18:03:05
jrsteensen/OpenHornet
https://api.github.com/repos/jrsteensen/OpenHornet
closed
Missing "Lip or Rings" on bottom through holes on the Radar Alt Case (all three)
Type: Bug/Obsolesce Category: MCAD Priority: Normal
### Discord Username Krikee ### Bug Summary The holes used to fasten the POT Mount (OH5A2A4-21) and the Radio Alt Stand off (OH5A2A4-15) parts to the Radio Alt Case seem to be missing the inner "lip or ring" on the bottom side of the Case. These rings can be seen in the holes in the side of the case. The issue being the screws simply pass through the holes on the bottom and do not retain the attached parts. ![Screenshot 2023-05-25 155450](https://github.com/jrsteensen/OpenHornet/assets/91638076/3b4f3b5f-c566-4957-b5b4-14e6976694a9) ### Expected Results Through holes on bottom should have a lip to allow the screw to support the attached part ### Actual Results Through holes on the bottom allow the screw to pass through the case and do not support the attached parts. Specifically the POT Mount and the Radio Alt Stand Off part. ### Screenshots/Images/Files ![Screenshot 2023-05-25 133356](https://github.com/jrsteensen/OpenHornet/assets/91638076/26be4778-91fa-47b8-aeed-d26b48339e5f) ### Applicable Part Numbers OH5A2A4-16 - RALT CASE; OH5A2A4-21 - RALT POT MOUNT; OH5A2A4-15 - RALT STEPPER STANDOFF ### Release Version 1.0.0-beta.1 ### Category Mechanical (Structure/Panels/Mechanisms) ### Applicable End Item(s) Right Console (RCON) ### Built to print? - [X] I built (or attempted to build) the part to the OpenHornet print without any deviations. - [ ] I am not building this part to the OH print. (List deviations in detail in the Miscellaneous Info text area below.) ### Miscellaneous Info _No response_
1.0
Missing "Lip or Rings" on bottom through holes on the Radar Alt Case (all three) - ### Discord Username Krikee ### Bug Summary The holes used to fasten the POT Mount (OH5A2A4-21) and the Radio Alt Stand off (OH5A2A4-15) parts to the Radio Alt Case seem to be missing the inner "lip or ring" on the bottom side of the Case. These rings can be seen in the holes in the side of the case. The issue being the screws simply pass through the holes on the bottom and do not retain the attached parts. ![Screenshot 2023-05-25 155450](https://github.com/jrsteensen/OpenHornet/assets/91638076/3b4f3b5f-c566-4957-b5b4-14e6976694a9) ### Expected Results Through holes on bottom should have a lip to allow the screw to support the attached part ### Actual Results Through holes on the bottom allow the screw to pass through the case and do not support the attached parts. Specifically the POT Mount and the Radio Alt Stand Off part. ### Screenshots/Images/Files ![Screenshot 2023-05-25 133356](https://github.com/jrsteensen/OpenHornet/assets/91638076/26be4778-91fa-47b8-aeed-d26b48339e5f) ### Applicable Part Numbers OH5A2A4-16 - RALT CASE; OH5A2A4-21 - RALT POT MOUNT; OH5A2A4-15 - RALT STEPPER STANDOFF ### Release Version 1.0.0-beta.1 ### Category Mechanical (Structure/Panels/Mechanisms) ### Applicable End Item(s) Right Console (RCON) ### Built to print? - [X] I built (or attempted to build) the part to the OpenHornet print without any deviations. - [ ] I am not building this part to the OH print. (List deviations in detail in the Miscellaneous Info text area below.) ### Miscellaneous Info _No response_
non_main
missing lip or rings on bottom through holes on the radar alt case all three discord username krikee bug summary the holes used to fasten the pot mount and the radio alt stand off parts to the radio alt case seem to be missing the inner lip or ring on the bottom side of the case these rings can be seen in the holes in the side of the case the issue being the screws simply pass through the holes on the bottom and do not retain the attached parts expected results through holes on bottom should have a lip to allow the screw to support the attached part actual results through holes on the bottom allow the screw to pass through the case and do not support the attached parts specifically the pot mount and the radio alt stand off part screenshots images files applicable part numbers ralt case ralt pot mount ralt stepper standoff release version beta category mechanical structure panels mechanisms applicable end item s right console rcon built to print i built or attempted to build the part to the openhornet print without any deviations i am not building this part to the oh print list deviations in detail in the miscellaneous info text area below miscellaneous info no response
0
134,529
18,471,926,128
IssuesEvent
2021-10-17 21:53:25
samq-ghdemo/JS-Demo
https://api.github.com/repos/samq-ghdemo/JS-Demo
opened
WS-2019-0493 (High) detected in handlebars-4.0.5.tgz
security vulnerability
## WS-2019-0493 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/nyc/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - istanbul-reports-1.0.0-alpha.8.tgz - :x: **handlebars-4.0.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-Demo/commit/210025573ddd44a379ebb23baeb6e2648a69b3d3">210025573ddd44a379ebb23baeb6e2648a69b3d3</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> handlebars before 3.0.8 and 4.x before 4.5.2 is vulnerable to Arbitrary Code Execution. The package's lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript in the system. <p>Publish Date: 2019-11-14 <p>URL: <a href=https://github.com/handlebars-lang/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e>WS-2019-0493</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p> <p>Release Date: 2019-11-14</p> <p>Fix Resolution: handlebars - 3.0.8,4.5.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-reports:1.0.0-alpha.8;handlebars:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 3.0.8,4.5.2"}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2019-0493","vulnerabilityDetails":"handlebars before 3.0.8 and 4.x before 4.5.2 is vulnerable to Arbitrary Code Execution. The package\u0027s lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript in the system.","vulnerabilityUrl":"https://github.com/handlebars-lang/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0493 (High) detected in handlebars-4.0.5.tgz - ## WS-2019-0493 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/nyc/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - istanbul-reports-1.0.0-alpha.8.tgz - :x: **handlebars-4.0.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-Demo/commit/210025573ddd44a379ebb23baeb6e2648a69b3d3">210025573ddd44a379ebb23baeb6e2648a69b3d3</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> handlebars before 3.0.8 and 4.x before 4.5.2 is vulnerable to Arbitrary Code Execution. The package's lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript in the system. <p>Publish Date: 2019-11-14 <p>URL: <a href=https://github.com/handlebars-lang/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e>WS-2019-0493</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p> <p>Release Date: 2019-11-14</p> <p>Fix Resolution: handlebars - 3.0.8,4.5.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-reports:1.0.0-alpha.8;handlebars:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 3.0.8,4.5.2"}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2019-0493","vulnerabilityDetails":"handlebars before 3.0.8 and 4.x before 4.5.2 is vulnerable to Arbitrary Code Execution. The package\u0027s lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript in the system.","vulnerabilityUrl":"https://github.com/handlebars-lang/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
ws high detected in handlebars tgz ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules nyc node modules handlebars package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz istanbul reports alpha tgz x handlebars tgz vulnerable library found in head commit a href found in base branch main vulnerability details handlebars before and x before is vulnerable to arbitrary code execution the package s lookup helper fails to properly validate templates allowing attackers to submit templates that execute arbitrary javascript in the system publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt if grunt contrib nodeunit nodeunit tap nyc istanbul reports alpha handlebars isminimumfixversionavailable true minimumfixversion handlebars basebranches vulnerabilityidentifier ws vulnerabilitydetails handlebars before and x before is vulnerable to arbitrary code execution the package lookup helper fails to properly validate templates allowing attackers to submit templates that execute arbitrary javascript in the system vulnerabilityurl
0
17,039
2,968,102,393
IssuesEvent
2015-07-13 08:02:15
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
[TEST-FAILURE] WanReplicationTest.EntryWithDefault_TTL_Replication_Issue
Team: Integration Type: Defect
``` java.lang.AssertionError: null at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at com.hazelcast.wan.WanReplicationTest$2.run(WanReplicationTest.java:235) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:740) at com.hazelcast.wan.WanReplicationTest.assertKeysIn(WanReplicationTest.java:233) at com.hazelcast.wan.WanReplicationTest.EntryWithDefault_TTL_Replication_Issue(WanReplicationTest.java:370) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-nightly/com.hazelcast$hazelcast/633/testReport/junit/com.hazelcast.wan/WanReplicationTest/EntryWithDefault_TTL_Replication_Issue/
1.0
[TEST-FAILURE] WanReplicationTest.EntryWithDefault_TTL_Replication_Issue - ``` java.lang.AssertionError: null at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at com.hazelcast.wan.WanReplicationTest$2.run(WanReplicationTest.java:235) at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:740) at com.hazelcast.wan.WanReplicationTest.assertKeysIn(WanReplicationTest.java:233) at com.hazelcast.wan.WanReplicationTest.EntryWithDefault_TTL_Replication_Issue(WanReplicationTest.java:370) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-nightly/com.hazelcast$hazelcast/633/testReport/junit/com.hazelcast.wan/WanReplicationTest/EntryWithDefault_TTL_Replication_Issue/
non_main
wanreplicationtest entrywithdefault ttl replication issue java lang assertionerror null at org junit assert fail assert java at org junit assert asserttrue assert java at org junit assert asserttrue assert java at com hazelcast wan wanreplicationtest run wanreplicationtest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast wan wanreplicationtest assertkeysin wanreplicationtest java at com hazelcast wan wanreplicationtest entrywithdefault ttl replication issue wanreplicationtest java
0
49,081
20,573,869,407
IssuesEvent
2022-03-04 00:55:55
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
closed
Alert when Postgres Has no synchronous standby
artifactory patroni ops medium priority ops and shared services
Patroni HA DB has occasionally lost it's synchronous standby and did not automatically re-initializing one. A restart of a replication member will reset the Sync Standby flag for a member. - [x] Alerting for this situation to be added - [x] Possible monitoring and auto-heal?
1.0
Alert when Postgres Has no synchronous standby - Patroni HA DB has occasionally lost it's synchronous standby and did not automatically re-initializing one. A restart of a replication member will reset the Sync Standby flag for a member. - [x] Alerting for this situation to be added - [x] Possible monitoring and auto-heal?
non_main
alert when postgres has no synchronous standby patroni ha db has occasionally lost it s synchronous standby and did not automatically re initializing one a restart of a replication member will reset the sync standby flag for a member alerting for this situation to be added possible monitoring and auto heal
0
597
4,097,245,323
IssuesEvent
2016-06-03 00:21:25
Particular/ServiceControl
https://api.github.com/repos/Particular/ServiceControl
closed
Deleting instance issues can be improved
Tag: Installer Tag: Maintainer Prio Type: Improvement
I just got the following dialog after trying to remove an instance: ![image](https://cloud.githubusercontent.com/assets/152998/14630120/bda79764-060b-11e6-915d-3be8fac1cbbd.png) Improvements: - The removal of the Logs folder *is mentioned twice* - It would be nice if I could open those folder by clicking a link - The message structure could be improved to make it less messy. Something like:: Logs directory: - Could not be deleted, please remove manually - Location: C:\ProgramData\Particular\ServiceControl\Particular.ServiceControl.MSMQ\Logs Database directory - Could not be deleted, please remove manually - Location: C:\ProgramData\Particular\ServiceControl\Particular.ServiceControl.MSMQ\DB
True
Deleting instance issues can be improved - I just got the following dialog after trying to remove an instance: ![image](https://cloud.githubusercontent.com/assets/152998/14630120/bda79764-060b-11e6-915d-3be8fac1cbbd.png) Improvements: - The removal of the Logs folder *is mentioned twice* - It would be nice if I could open those folder by clicking a link - The message structure could be improved to make it less messy. Something like:: Logs directory: - Could not be deleted, please remove manually - Location: C:\ProgramData\Particular\ServiceControl\Particular.ServiceControl.MSMQ\Logs Database directory - Could not be deleted, please remove manually - Location: C:\ProgramData\Particular\ServiceControl\Particular.ServiceControl.MSMQ\DB
main
deleting instance issues can be improved i just got the following dialog after trying to remove an instance improvements the removal of the logs folder is mentioned twice it would be nice if i could open those folder by clicking a link the message structure could be improved to make it less messy something like logs directory could not be deleted please remove manually location c programdata particular servicecontrol particular servicecontrol msmq logs database directory could not be deleted please remove manually location c programdata particular servicecontrol particular servicecontrol msmq db
1
397,372
11,727,567,143
IssuesEvent
2020-03-10 16:08:13
kubernetes-sigs/kind
https://api.github.com/repos/kubernetes-sigs/kind
closed
Using flannel as CNI no longer works
kind/external priority/important-soon
**What happened**: After installing Flannel as my CNI, pods fail to start with the error `Failed to create pod sandbox: rpc error: code = Unknown desc = failed to setup network for sandbox "40a61cb16293c414497736143f53da331ef0dca2236e223f3057bd930d51c1c6": failed to find plugin "flannel" in path [/opt/cni/bin]` **What you expected to happen**: Pods work out of the box as before! **How to reproduce it (as minimally and precisely as possible)**: Using 0.7.0, create a cluster with `networking.disableDefaultCNI: true`. Install the Flannel CNI [yaml](https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yml). See that CoreDNS pods in `kube-system` fail to start. **Anything else we need to know?**: It works in 0.6.1, it looks like https://github.com/kubernetes-sigs/kind/commit/281a20c36c91da3347bc514512df10229d053c50 is what broke it. Normally Flannel assumes that the flannel CNI binary is included on the host, since it is part of the `kubernetes-cni` package and the default https://github.com/containernetworking/plugins repo.
1.0
Using flannel as CNI no longer works - **What happened**: After installing Flannel as my CNI, pods fail to start with the error `Failed to create pod sandbox: rpc error: code = Unknown desc = failed to setup network for sandbox "40a61cb16293c414497736143f53da331ef0dca2236e223f3057bd930d51c1c6": failed to find plugin "flannel" in path [/opt/cni/bin]` **What you expected to happen**: Pods work out of the box as before! **How to reproduce it (as minimally and precisely as possible)**: Using 0.7.0, create a cluster with `networking.disableDefaultCNI: true`. Install the Flannel CNI [yaml](https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yml). See that CoreDNS pods in `kube-system` fail to start. **Anything else we need to know?**: It works in 0.6.1, it looks like https://github.com/kubernetes-sigs/kind/commit/281a20c36c91da3347bc514512df10229d053c50 is what broke it. Normally Flannel assumes that the flannel CNI binary is included on the host, since it is part of the `kubernetes-cni` package and the default https://github.com/containernetworking/plugins repo.
non_main
using flannel as cni no longer works what happened after installing flannel as my cni pods fail to start with the error failed to create pod sandbox rpc error code unknown desc failed to setup network for sandbox failed to find plugin flannel in path what you expected to happen pods work out of the box as before how to reproduce it as minimally and precisely as possible using create a cluster with networking disabledefaultcni true install the flannel cni see that coredns pods in kube system fail to start anything else we need to know it works in it looks like is what broke it normally flannel assumes that the flannel cni binary is included on the host since it is part of the kubernetes cni package and the default repo
0
642
4,157,822,349
IssuesEvent
2016-06-16 22:49:32
coniks-sys/coniks-ref-implementation
https://api.github.com/repos/coniks-sys/coniks-ref-implementation
opened
Make functions that serialize nodes instance methods
maintainability server
Refactor all functions in *ServerUtils.java* that convert tree nodes ans STRs into byte arrays and make them instance methods for each type.
True
Make functions that serialize nodes instance methods - Refactor all functions in *ServerUtils.java* that convert tree nodes ans STRs into byte arrays and make them instance methods for each type.
main
make functions that serialize nodes instance methods refactor all functions in serverutils java that convert tree nodes ans strs into byte arrays and make them instance methods for each type
1
3,661
14,942,818,775
IssuesEvent
2021-01-25 21:55:16
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
DataTable - How to make rows disabled in order not to be selected when clicking on the SelectAll checkbox?
status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬 type: question ❓
<!-- Hi there! 👋 Hope everything is going okay using projects from the Carbon Design System. It looks like you might have a question about our work, so we wanted to share a couple resources that you could use if you haven't tried them yet 🙂. If you're an IBMer, we have a couple of Slack channels available across all IBM Workspaces: - #carbon-design-system for questions about the Design System - #carbon-components for questions about component styles - #carbon-react for questions about our React components If these resources don't work out, help us out by filling out a couple of details below! --> ## What package(s) are you using? <!-- Add an x in one of the options below, for example: - [x] package name --> - [ ] `carbon-components` - [x] `carbon-components-react` ## Summary I disabled the `<TableSelectRow>` but when clicking on SelectAll these rows are selected, and I need them to be unselected. ## Relevant information Tried to apply the disabled attribute to `<TableRow>` but it didn't help! **Note**: I have a menu near the name of the row with "Info" option, so disabling the row might block this functionality -which is not wanted- `<TableBody> {rows.map(row => ( <TableRow key={row.id}> <TableSelectRow {...getSelectionProps({ row, onClick: event => { //console.log(row.id, event.target.checked); this.selectRule(row.id, event.target.checked) } }) } disabled={this.state.unattachedRulesTables[row.id] && this.state.unattachedRulesTables[row.id].length > 0}/> {row.cells.map(cell => ( <TableCell key={cell.id}>{cell.value}</TableCell> ))} </TableRow> ))} </TableBody>` <!-- Provide as much useful information as you can -->
True
DataTable - How to make rows disabled in order not to be selected when clicking on the SelectAll checkbox? - <!-- Hi there! 👋 Hope everything is going okay using projects from the Carbon Design System. It looks like you might have a question about our work, so we wanted to share a couple resources that you could use if you haven't tried them yet 🙂. If you're an IBMer, we have a couple of Slack channels available across all IBM Workspaces: - #carbon-design-system for questions about the Design System - #carbon-components for questions about component styles - #carbon-react for questions about our React components If these resources don't work out, help us out by filling out a couple of details below! --> ## What package(s) are you using? <!-- Add an x in one of the options below, for example: - [x] package name --> - [ ] `carbon-components` - [x] `carbon-components-react` ## Summary I disabled the `<TableSelectRow>` but when clicking on SelectAll these rows are selected, and I need them to be unselected. ## Relevant information Tried to apply the disabled attribute to `<TableRow>` but it didn't help! **Note**: I have a menu near the name of the row with "Info" option, so disabling the row might block this functionality -which is not wanted- `<TableBody> {rows.map(row => ( <TableRow key={row.id}> <TableSelectRow {...getSelectionProps({ row, onClick: event => { //console.log(row.id, event.target.checked); this.selectRule(row.id, event.target.checked) } }) } disabled={this.state.unattachedRulesTables[row.id] && this.state.unattachedRulesTables[row.id].length > 0}/> {row.cells.map(cell => ( <TableCell key={cell.id}>{cell.value}</TableCell> ))} </TableRow> ))} </TableBody>` <!-- Provide as much useful information as you can -->
main
datatable how to make rows disabled in order not to be selected when clicking on the selectall checkbox hi there 👋 hope everything is going okay using projects from the carbon design system it looks like you might have a question about our work so we wanted to share a couple resources that you could use if you haven t tried them yet 🙂 if you re an ibmer we have a couple of slack channels available across all ibm workspaces carbon design system for questions about the design system carbon components for questions about component styles carbon react for questions about our react components if these resources don t work out help us out by filling out a couple of details below what package s are you using add an x in one of the options below for example package name carbon components carbon components react summary i disabled the but when clicking on selectall these rows are selected and i need them to be unselected relevant information tried to apply the disabled attribute to but it didn t help note i have a menu near the name of the row with info option so disabling the row might block this functionality which is not wanted rows map row tableselectrow getselectionprops row onclick event console log row id event target checked this selectrule row id event target checked disabled this state unattachedrulestables this state unattachedrulestables length row cells map cell cell value
1
5,466
27,330,654,803
IssuesEvent
2023-02-25 15:33:53
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[QUESTION] How does "sortBy": "most-used" work ?
🤷‍♂️ Question 👤 Awaiting Maintainer Response
### Question I've set `"sortBy": "most-used",` for all of the sections on the dash, but none of the items inside of the sections are changing positions. Does something else need to be set for sorting by most-used ? ### Category Configuration ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
True
[QUESTION] How does "sortBy": "most-used" work ? - ### Question I've set `"sortBy": "most-used",` for all of the sections on the dash, but none of the items inside of the sections are changing positions. Does something else need to be set for sorting by most-used ? ### Category Configuration ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
main
how does sortby most used work question i ve set sortby most used for all of the sections on the dash but none of the items inside of the sections are changing positions does something else need to be set for sorting by most used category configuration please tick the boxes you are using a version of dashy check the first two digits of the version number you ve checked that this you ve checked the and guide you agree to the
1
60,652
17,023,483,458
IssuesEvent
2021-07-03 02:15:49
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
It seems changesets without a bounding box do not show up in a user's list of edits
Component: website Priority: minor Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 12.39pm, Friday, 25th September 2009]** When you browse this user's list of edits page (http://www.openstreetmap.org/user/maning/edits), you don't see this changeset of his: http://www.openstreetmap.org/browse/changeset/2611259 (though this changeset can be reached by following the next/previous changeset links) It seems that it's not listed because there was no bounding box (the changeset only deleted a relation).
1.0
It seems changesets without a bounding box do not show up in a user's list of edits - **[Submitted to the original trac issue database at 12.39pm, Friday, 25th September 2009]** When you browse this user's list of edits page (http://www.openstreetmap.org/user/maning/edits), you don't see this changeset of his: http://www.openstreetmap.org/browse/changeset/2611259 (though this changeset can be reached by following the next/previous changeset links) It seems that it's not listed because there was no bounding box (the changeset only deleted a relation).
non_main
it seems changesets without a bounding box do not show up in a user s list of edits when you browse this user s list of edits page you don t see this changeset of his though this changeset can be reached by following the next previous changeset links it seems that it s not listed because there was no bounding box the changeset only deleted a relation
0
2,055
6,972,730,941
IssuesEvent
2017-12-11 17:59:28
CyberReboot/vent
https://api.github.com/repos/CyberReboot/vent
closed
build should be optional, if image with the correct tag already exists
area/administration area/plugins area/quality/maintainability
useful for no internet scenarios
True
build should be optional, if image with the correct tag already exists - useful for no internet scenarios
main
build should be optional if image with the correct tag already exists useful for no internet scenarios
1