Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
236,993
7,754,657,106
IssuesEvent
2018-05-31 07:35:36
FACG4/FACGTracker
https://api.github.com/repos/FACG4/FACGTracker
closed
Add feedback
priority-2 user-story
## Acceptance criteria : - [x] Authentication - [x] DB query for : * delete feedback * add feedback - [x] feedback hbs and route
1.0
Add feedback - ## Acceptance criteria : - [x] Authentication - [x] DB query for : * delete feedback * add feedback - [x] feedback hbs and route
non_process
add feedback acceptance criteria authentication db query for delete feedback add feedback feedback hbs and route
0
339,764
30,472,933,551
IssuesEvent
2023-07-17 14:41:20
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix discrete_fourier_transform.test_numpy_fft
NumPy Frontend Sub Task Failing Test
| | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix discrete_fourier_transform.test_numpy_fft - | | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5576979617"><img src=https://img.shields.io/badge/-success-success></a>
non_process
fix discrete fourier transform test numpy fft jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
0
17,959
23,963,436,705
IssuesEvent
2022-09-12 21:28:24
elastic/beats
https://api.github.com/repos/elastic/beats
opened
[libbeat] add_cloud_metadata - startup blocked by AWS IMSDv2 token fetch
bug libbeat :Processors
The AWS provider in the `add_cloud_metadata` processor was modified to support IMSDv2 in https://github.com/elastic/beats/pull/28285. There are two issues affecting the implementation. 1. Fetching of the IMSDv2 token blocks the startup of the processor. The processor was designed to try fetching metadata from all enabled providers (e.g. aws, gcp, openstack) in parallel and asynchronously from the main thread. This allows the Beat to startup without being blocked on Cloud provider detection logic. \ The implementation fetches the IMSDv2 token in a non-async manner such that it blocks startup and does not happen in parallel with all other provider implementations. In non-AWS environments the Beat blocks until a HTTP timeout is reached. \ The processor needs to be modified to fetch the IMSDv2 token asynchronously. Some changes will be needed to allow chaining the request to fetch the token and use the token in the AWS metadata service requests. 2. It logs warnings for something that is expected to fail when the Beat runs outside of AWS. This should not be logged as a warning. It should be handled in the same manner as a non-response is handled by all other provider implementations. If the error is returned in the result from the provider then the processor will [log](https://github.com/elastic/beats/blob/25786cdda70b31cb1738373265bf3a0f3dec76f6/libbeat/processors/add_cloud_metadata/providers.go#L167-L168) the disposition at the debug level. > {"log.level":"warn","@timestamp":"2022-09-12T17:09:41.120-0400","log.logger":"add_cloud_metadata","log.origin":{"file.name":"add_cloud_metadata/provider_aws_ec2.go","file.line":81},"message":"read token request for getting IMDSv2 token returns empty: Put \"http://169.254.169.254/latest/api/token\": context deadline exceeded (Client.Timeout exceeded while awaiting headers). No token in the metadata request will be used.","service.name":"filebeat","ecs.version":"1.6.0"}
1.0
[libbeat] add_cloud_metadata - startup blocked by AWS IMSDv2 token fetch - The AWS provider in the `add_cloud_metadata` processor was modified to support IMSDv2 in https://github.com/elastic/beats/pull/28285. There are two issues affecting the implementation. 1. Fetching of the IMSDv2 token blocks the startup of the processor. The processor was designed to try fetching metadata from all enabled providers (e.g. aws, gcp, openstack) in parallel and asynchronously from the main thread. This allows the Beat to startup without being blocked on Cloud provider detection logic. \ The implementation fetches the IMSDv2 token in a non-async manner such that it blocks startup and does not happen in parallel with all other provider implementations. In non-AWS environments the Beat blocks until a HTTP timeout is reached. \ The processor needs to be modified to fetch the IMSDv2 token asynchronously. Some changes will be needed to allow chaining the request to fetch the token and use the token in the AWS metadata service requests. 2. It logs warnings for something that is expected to fail when the Beat runs outside of AWS. This should not be logged as a warning. It should be handled in the same manner as a non-response is handled by all other provider implementations. If the error is returned in the result from the provider then the processor will [log](https://github.com/elastic/beats/blob/25786cdda70b31cb1738373265bf3a0f3dec76f6/libbeat/processors/add_cloud_metadata/providers.go#L167-L168) the disposition at the debug level. > {"log.level":"warn","@timestamp":"2022-09-12T17:09:41.120-0400","log.logger":"add_cloud_metadata","log.origin":{"file.name":"add_cloud_metadata/provider_aws_ec2.go","file.line":81},"message":"read token request for getting IMDSv2 token returns empty: Put \"http://169.254.169.254/latest/api/token\": context deadline exceeded (Client.Timeout exceeded while awaiting headers). No token in the metadata request will be used.","service.name":"filebeat","ecs.version":"1.6.0"}
process
add cloud metadata startup blocked by aws token fetch the aws provider in the add cloud metadata processor was modified to support in there are two issues affecting the implementation fetching of the token blocks the startup of the processor the processor was designed to try fetching metadata from all enabled providers e g aws gcp openstack in parallel and asynchronously from the main thread this allows the beat to startup without being blocked on cloud provider detection logic the implementation fetches the token in a non async manner such that it blocks startup and does not happen in parallel with all other provider implementations in non aws environments the beat blocks until a http timeout is reached the processor needs to be modified to fetch the token asynchronously some changes will be needed to allow chaining the request to fetch the token and use the token in the aws metadata service requests it logs warnings for something that is expected to fail when the beat runs outside of aws this should not be logged as a warning it should be handled in the same manner as a non response is handled by all other provider implementations if the error is returned in the result from the provider then the processor will the disposition at the debug level log level warn timestamp log logger add cloud metadata log origin file name add cloud metadata provider aws go file line message read token request for getting token returns empty put context deadline exceeded client timeout exceeded while awaiting headers no token in the metadata request will be used service name filebeat ecs version
1
246,723
7,895,616,836
IssuesEvent
2018-06-29 04:33:55
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
The Xdmf reader shows all the sub-grids for a collection instead of just the collection.
Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 2 - Minor Irritation Support Group: Any bug version: 2.0.0
If a file has a grid that is a collection of grids it currently returns all the grids that make up the collection instead of a single grid that is the collection of grids. Both Dave Pugmire of ORNL and Jorge Pita of ARAMCO have run into this problem. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Eric Brugger Original creation: 01/12/2011 05:12 pm Original update: 01/12/2011 06:12 pm Ticket number: 557
1.0
The Xdmf reader shows all the sub-grids for a collection instead of just the collection. - If a file has a grid that is a collection of grids it currently returns all the grids that make up the collection instead of a single grid that is the collection of grids. Both Dave Pugmire of ORNL and Jorge Pita of ARAMCO have run into this problem. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Eric Brugger Original creation: 01/12/2011 05:12 pm Original update: 01/12/2011 06:12 pm Ticket number: 557
non_process
the xdmf reader shows all the sub grids for a collection instead of just the collection if a file has a grid that is a collection of grids it currently returns all the grids that make up the collection instead of a single grid that is the collection of grids both dave pugmire of ornl and jorge pita of aramco have run into this problem redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author eric brugger original creation pm original update pm ticket number
0
13,947
16,723,336,266
IssuesEvent
2021-06-10 09:56:29
hochschule-darmstadt/openartbrowser
https://api.github.com/repos/hochschule-darmstadt/openartbrowser
closed
Blocklist functionality in etl
etl process medium priority
**Reason (Why?)** If some categories seem inappropriate, there should be a functionality to filter them from the dataset. **Solution (What?)** Implementation of a blocklist, which is applied to the crawling process in a performant way. The blocklist should be a json array, containing all QIDs of block-listeted entities (all types possible). As a result, the entity and all references to it should not appear in the dataset. **Acceptance criteria** Block-listed entities no longer appear in the dataset.
1.0
Blocklist functionality in etl - **Reason (Why?)** If some categories seem inappropriate, there should be a functionality to filter them from the dataset. **Solution (What?)** Implementation of a blocklist, which is applied to the crawling process in a performant way. The blocklist should be a json array, containing all QIDs of block-listeted entities (all types possible). As a result, the entity and all references to it should not appear in the dataset. **Acceptance criteria** Block-listed entities no longer appear in the dataset.
process
blocklist functionality in etl reason why if some categories seem inappropriate there should be a functionality to filter them from the dataset solution what implementation of a blocklist which is applied to the crawling process in a performant way the blocklist should be a json array containing all qids of block listeted entities all types possible as a result the entity and all references to it should not appear in the dataset acceptance criteria block listed entities no longer appear in the dataset
1
2,408
5,193,320,973
IssuesEvent
2017-01-22 18:10:03
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[subtitles] [eng] #RDLS14 - FRANCE INTER, TRUMP ET LA CHINE, FILLON ET INTERNET, LE CAC40 SE GAVE,
Language: English Process: [6] Approved
# Video title #RDLS14 - FRANCE INTER, TRUMP ET LA CHINE, FILLON ET INTERNET, LE CAC40 SE GAVE, # URL https://www.youtube.com/watch?v=xSxMAU-DZsc # Youtube subtitles language Anglais # Duration 19:13 # Subtitles URL https://www.youtube.com/timedtext_editor?bl=vmp&lang=en&action_mde_edit_form=1&ref=player&tab=captions&v=xSxMAU-DZsc&ui=hd
1.0
[subtitles] [eng] #RDLS14 - FRANCE INTER, TRUMP ET LA CHINE, FILLON ET INTERNET, LE CAC40 SE GAVE, - # Video title #RDLS14 - FRANCE INTER, TRUMP ET LA CHINE, FILLON ET INTERNET, LE CAC40 SE GAVE, # URL https://www.youtube.com/watch?v=xSxMAU-DZsc # Youtube subtitles language Anglais # Duration 19:13 # Subtitles URL https://www.youtube.com/timedtext_editor?bl=vmp&lang=en&action_mde_edit_form=1&ref=player&tab=captions&v=xSxMAU-DZsc&ui=hd
process
france inter trump et la chine fillon et internet le se gave video title france inter trump et la chine fillon et internet le se gave url youtube subtitles language anglais duration subtitles url
1
131,668
5,163,637,150
IssuesEvent
2017-01-17 07:43:21
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
TypeError: Cannot read property 'nID' of undefined api/messages/index.controller
active bug hi priority _central-js
2016-10-18_23:48:32.376: TypeError: Cannot read property 'nID' of undefined at module.exports.postServiceMessage (/sybase/central-js/server/api/messages/index.controller.js:138:28) at Layer.handle [as handle_request](/sybase/central-js/node_modules/express/lib/router/layer.js:95:5) at next (/sybase/central-js/node_modules/express/lib/router/route.js:131:13) at Route.dispatch (/sybase/central-js/node_modules/express/lib/router/route.js:112:3) at Layer.handle [as handle_request](/sybase/central-js/node_modules/express/lib/router/layer.js:95:5) at /sybase/central-js/node_modules/express/lib/router/index.js:277:22 at Function.process_params (/sybase/central-js/node_modules/express/lib/router/index.js:330:12) at next (/sybase/central-js/node_modules/express/lib/router/index.js:271:10) at Function.handle (/sybase/central-js/node_modules/express/lib/router/index.js:176:3) at router (/sybase/central-js/node_modules/express/lib/router/index.js:46:12)
1.0
TypeError: Cannot read property 'nID' of undefined api/messages/index.controller - 2016-10-18_23:48:32.376: TypeError: Cannot read property 'nID' of undefined at module.exports.postServiceMessage (/sybase/central-js/server/api/messages/index.controller.js:138:28) at Layer.handle [as handle_request](/sybase/central-js/node_modules/express/lib/router/layer.js:95:5) at next (/sybase/central-js/node_modules/express/lib/router/route.js:131:13) at Route.dispatch (/sybase/central-js/node_modules/express/lib/router/route.js:112:3) at Layer.handle [as handle_request](/sybase/central-js/node_modules/express/lib/router/layer.js:95:5) at /sybase/central-js/node_modules/express/lib/router/index.js:277:22 at Function.process_params (/sybase/central-js/node_modules/express/lib/router/index.js:330:12) at next (/sybase/central-js/node_modules/express/lib/router/index.js:271:10) at Function.handle (/sybase/central-js/node_modules/express/lib/router/index.js:176:3) at router (/sybase/central-js/node_modules/express/lib/router/index.js:46:12)
non_process
typeerror cannot read property nid of undefined api messages index controller typeerror cannot read property nid of undefined at module exports postservicemessage sybase central js server api messages index controller js at layer handle sybase central js node modules express lib router layer js at next sybase central js node modules express lib router route js at route dispatch sybase central js node modules express lib router route js at layer handle sybase central js node modules express lib router layer js at sybase central js node modules express lib router index js at function process params sybase central js node modules express lib router index js at next sybase central js node modules express lib router index js at function handle sybase central js node modules express lib router index js at router sybase central js node modules express lib router index js
0
85,241
3,688,257,847
IssuesEvent
2016-02-25 12:03:08
RestComm/mediaserver
https://api.github.com/repos/RestComm/mediaserver
closed
Pooled connections cannot handle multiple WebRTC outbound calls
bug High-Priority in progress
An issue was identified where RTP connections cannot do multiple WebRTC outbound calls. Once the connection is pooled (after first call), following attempts to do outbound WebRTC call will result in the following error: 13:58:08,325 INFO [RtpChannel] Can not connect to remote address , please check that you are not using local address - 127.0.0.X to connect to remote 13:58:08,327 ERROR [RtpChannel] java.nio.channels.ClosedChannelException at sun.nio.ch.DatagramChannelImpl.ensureOpenAndUnconnected(DatagramChannelImpl.java:719) at sun.nio.ch.DatagramChannelImpl.connect(DatagramChannelImpl.java:732) at org.mobicents.media.server.impl.rtp.RtpChannel.setRemotePeer(RtpChannel.java:369) at org.mobicents.media.server.impl.rtp.channels.MediaChannel.connectRtp(MediaChannel.java:506) at org.mobicents.media.server.impl.rtp.channels.MediaChannel.connectRtp(MediaChannel.java:527) at org.mobicents.media.core.connections.RtpConnectionImpl.setupAudioChannelOutbound(RtpConnectionImpl.java:387) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherPartyOutboundCall(RtpConnectionImpl.java:291) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherParty(RtpConnectionImpl.java:206) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherParty(RtpConnectionImpl.java:163) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherParty(RtpConnectionImpl.java:171) at org.mobicents.media.server.mgcp.controller.MgcpConnection.setOtherParty(MgcpConnection.java:128) at org.mobicents.media.server.mgcp.tx.cmd.ModifyConnectionCmd$Modifier.perform(ModifyConnectionCmd.java:136) at org.mobicents.media.server.scheduler.Task.run(Task.java:122) at org.mobicents.media.server.scheduler.PriorityQueueScheduler$WorkerThread.run(PriorityQueueScheduler.java:424)
1.0
Pooled connections cannot handle multiple WebRTC outbound calls - An issue was identified where RTP connections cannot do multiple WebRTC outbound calls. Once the connection is pooled (after first call), following attempts to do outbound WebRTC call will result in the following error: 13:58:08,325 INFO [RtpChannel] Can not connect to remote address , please check that you are not using local address - 127.0.0.X to connect to remote 13:58:08,327 ERROR [RtpChannel] java.nio.channels.ClosedChannelException at sun.nio.ch.DatagramChannelImpl.ensureOpenAndUnconnected(DatagramChannelImpl.java:719) at sun.nio.ch.DatagramChannelImpl.connect(DatagramChannelImpl.java:732) at org.mobicents.media.server.impl.rtp.RtpChannel.setRemotePeer(RtpChannel.java:369) at org.mobicents.media.server.impl.rtp.channels.MediaChannel.connectRtp(MediaChannel.java:506) at org.mobicents.media.server.impl.rtp.channels.MediaChannel.connectRtp(MediaChannel.java:527) at org.mobicents.media.core.connections.RtpConnectionImpl.setupAudioChannelOutbound(RtpConnectionImpl.java:387) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherPartyOutboundCall(RtpConnectionImpl.java:291) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherParty(RtpConnectionImpl.java:206) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherParty(RtpConnectionImpl.java:163) at org.mobicents.media.core.connections.RtpConnectionImpl.setOtherParty(RtpConnectionImpl.java:171) at org.mobicents.media.server.mgcp.controller.MgcpConnection.setOtherParty(MgcpConnection.java:128) at org.mobicents.media.server.mgcp.tx.cmd.ModifyConnectionCmd$Modifier.perform(ModifyConnectionCmd.java:136) at org.mobicents.media.server.scheduler.Task.run(Task.java:122) at org.mobicents.media.server.scheduler.PriorityQueueScheduler$WorkerThread.run(PriorityQueueScheduler.java:424)
non_process
pooled connections cannot handle multiple webrtc outbound calls an issue was identified where rtp connections cannot do multiple webrtc outbound calls once the connection is pooled after first call following attempts to do outbound webrtc call will result in the following error info can not connect to remote address please check that you are not using local address x to connect to remote error java nio channels closedchannelexception at sun nio ch datagramchannelimpl ensureopenandunconnected datagramchannelimpl java at sun nio ch datagramchannelimpl connect datagramchannelimpl java at org mobicents media server impl rtp rtpchannel setremotepeer rtpchannel java at org mobicents media server impl rtp channels mediachannel connectrtp mediachannel java at org mobicents media server impl rtp channels mediachannel connectrtp mediachannel java at org mobicents media core connections rtpconnectionimpl setupaudiochanneloutbound rtpconnectionimpl java at org mobicents media core connections rtpconnectionimpl setotherpartyoutboundcall rtpconnectionimpl java at org mobicents media core connections rtpconnectionimpl setotherparty rtpconnectionimpl java at org mobicents media core connections rtpconnectionimpl setotherparty rtpconnectionimpl java at org mobicents media core connections rtpconnectionimpl setotherparty rtpconnectionimpl java at org mobicents media server mgcp controller mgcpconnection setotherparty mgcpconnection java at org mobicents media server mgcp tx cmd modifyconnectioncmd modifier perform modifyconnectioncmd java at org mobicents media server scheduler task run task java at org mobicents media server scheduler priorityqueuescheduler workerthread run priorityqueuescheduler java
0
118,188
11,964,072,295
IssuesEvent
2020-04-05 18:08:25
aws/amazon-chime-sdk-js
https://api.github.com/repos/aws/amazon-chime-sdk-js
opened
Document Supported Browser Check
documentation
### What are you trying to do? I am trying to display a message to the user when their browser is not supported. ### How can the documentation be improved to help your use case? What is the recommended way to check browser support? I'm currently checking that the browser supports webRTC (RTCPeerConnection, webkitRTCPeerConnection, mozRTCPeerConnection, RTCIceGatherer), but this is not comprehensive. Some earlier versions of Edge pass this check but are still not supported. In the API docs, I noticed there is a DefaultBrowserBehavior class with an isSupported() method; This isn't used in the demo - is this what I should be using? ### What documentation have you looked at so far? README.md, Amazon Chime SDK Developer Guide, Getting Started, API Overview, code demos, API docs, etc.
1.0
Document Supported Browser Check - ### What are you trying to do? I am trying to display a message to the user when their browser is not supported. ### How can the documentation be improved to help your use case? What is the recommended way to check browser support? I'm currently checking that the browser supports webRTC (RTCPeerConnection, webkitRTCPeerConnection, mozRTCPeerConnection, RTCIceGatherer), but this is not comprehensive. Some earlier versions of Edge pass this check but are still not supported. In the API docs, I noticed there is a DefaultBrowserBehavior class with an isSupported() method; This isn't used in the demo - is this what I should be using? ### What documentation have you looked at so far? README.md, Amazon Chime SDK Developer Guide, Getting Started, API Overview, code demos, API docs, etc.
non_process
document supported browser check what are you trying to do i am trying to display a message to the user when their browser is not supported how can the documentation be improved to help your use case what is the recommended way to check browser support i m currently checking that the browser supports webrtc rtcpeerconnection webkitrtcpeerconnection mozrtcpeerconnection rtcicegatherer but this is not comprehensive some earlier versions of edge pass this check but are still not supported in the api docs i noticed there is a defaultbrowserbehavior class with an issupported method this isn t used in the demo is this what i should be using what documentation have you looked at so far readme md amazon chime sdk developer guide getting started api overview code demos api docs etc
0
7,354
10,483,672,928
IssuesEvent
2019-09-24 14:20:06
toggl/mobileapp
https://api.github.com/repos/toggl/mobileapp
opened
Figure out production/stating Firebase configuration
process
Currently we only one Firebase project for our apps's configurations. Ideally, we should have different api keys for production and staging environments. Using the current configuration, I couldn't find a way to generate a new api key that would isolate production apps from staging/debug apps. The way I see we have two options: 1 - Contact Firebase support and ask to make sure there isn't a way to create an isolated api key without creating another firebase project. 2 - Create another firebase project for staging (and use the current one as production). Pinging @amulware ➿ and _🔨 (loop & decisions).
1.0
Figure out production/stating Firebase configuration - Currently we only one Firebase project for our apps's configurations. Ideally, we should have different api keys for production and staging environments. Using the current configuration, I couldn't find a way to generate a new api key that would isolate production apps from staging/debug apps. The way I see we have two options: 1 - Contact Firebase support and ask to make sure there isn't a way to create an isolated api key without creating another firebase project. 2 - Create another firebase project for staging (and use the current one as production). Pinging @amulware ➿ and _🔨 (loop & decisions).
process
figure out production stating firebase configuration currently we only one firebase project for our apps s configurations ideally we should have different api keys for production and staging environments using the current configuration i couldn t find a way to generate a new api key that would isolate production apps from staging debug apps the way i see we have two options contact firebase support and ask to make sure there isn t a way to create an isolated api key without creating another firebase project create another firebase project for staging and use the current one as production pinging amulware ➿ and 🔨 loop decisions
1
9,550
12,513,731,427
IssuesEvent
2020-06-03 02:53:46
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Sample has errors due to Powershell commands
Pri3 automation/svc cxp process-automation/subsvc product-issue triaged
I am following the steps listed in this document, by testing the sample code: https://docs.microsoft.com/en-us/azure/automation/automation-webhooks I have successfully created the webhook, runbook and JSON request in powershell. I can verify this as the output in the RunBook shows: **Header has required information** **Authenticating to Azure with service principal and certificate** **Get connection asset: AzureRunAsConnection** **Authenticating to Azure with service principal.** **Starting Machine-name-here** I have replaced my actual machine name with "machine-name-here" for security reasons. My input is (again, I redact the Azure webhook, storage group and machine name for security reasons): ``` $uri = "https://XXXXXXX" $vms = @( @{ Name="Machine-name-here";ResourceGroup="storage-group-name"} ) $body = ConvertTo-Json -InputObject $vms $header = @{ message="StartedbyUser"} $response = Invoke-WebRequest -Method Post -Uri $uri -Body $body -Headers $header $jobid = (ConvertFrom-Json ($response.Content)).jobids[0 ``` This shows as completed in the runbook, but in the Errors section I see this: ``` Start-AzAccount : The term 'Start-AzAccount' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. At line:35 char:13 + Start-AzAccount -ServicePrincipal -Tenant $Conn.TenantID ... + ~~~~~~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (Start-AzAccount:String) [], CommandNotFoundException + FullyQualifiedErrorId : CommandNotFoundException ``` ``` Start-AzVM : The term 'Start-AzVM' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. At line:42 char:13 + Start-AzVM -Name $vm.Name -ResourceGroup $vm.ResourceGrou ... + ~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (Start-AzVM:String) [], CommandNotFoundException + FullyQualifiedErrorId : CommandNotFoundException ``` I have tried running this from multiple locations, and I receive the same error each time. I tried running this from Azure Cloud shell and received the same outcome. I have added Az.Compute before running the task on my local machine, and in cloud shell, only to have the same outcome. I tried adding in "Import-Module AzureRm" as an extra line to the runbook immediately prior to the Add-AzAccount line, but this requires user intervention and this script is run in a space which does not allow user intervention. What is required to allow these commands to complete? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7a6394c7-9bef-b8f8-ffd6-9d9d8e2daa07 * Version Independent ID: 5ffa20a2-436c-2726-dc57-9d3b49f9ca39 * Content: [Start an Azure Automation runbook from a webhook](https://docs.microsoft.com/en-us/azure/automation/automation-webhooks) * Content Source: [articles/automation/automation-webhooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-webhooks.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Sample has errors due to Powershell commands - I am following the steps listed in this document, by testing the sample code: https://docs.microsoft.com/en-us/azure/automation/automation-webhooks I have successfully created the webhook, runbook and JSON request in powershell. I can verify this as the output in the RunBook shows: **Header has required information** **Authenticating to Azure with service principal and certificate** **Get connection asset: AzureRunAsConnection** **Authenticating to Azure with service principal.** **Starting Machine-name-here** I have replaced my actual machine name with "machine-name-here" for security reasons. My input is (again, I redact the Azure webhook, storage group and machine name for security reasons): ``` $uri = "https://XXXXXXX" $vms = @( @{ Name="Machine-name-here";ResourceGroup="storage-group-name"} ) $body = ConvertTo-Json -InputObject $vms $header = @{ message="StartedbyUser"} $response = Invoke-WebRequest -Method Post -Uri $uri -Body $body -Headers $header $jobid = (ConvertFrom-Json ($response.Content)).jobids[0 ``` This shows as completed in the runbook, but in the Errors section I see this: ``` Start-AzAccount : The term 'Start-AzAccount' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. At line:35 char:13 + Start-AzAccount -ServicePrincipal -Tenant $Conn.TenantID ... + ~~~~~~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (Start-AzAccount:String) [], CommandNotFoundException + FullyQualifiedErrorId : CommandNotFoundException ``` ``` Start-AzVM : The term 'Start-AzVM' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. At line:42 char:13 + Start-AzVM -Name $vm.Name -ResourceGroup $vm.ResourceGrou ... + ~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (Start-AzVM:String) [], CommandNotFoundException + FullyQualifiedErrorId : CommandNotFoundException ``` I have tried running this from multiple locations, and I receive the same error each time. I tried running this from Azure Cloud shell and received the same outcome. I have added Az.Compute before running the task on my local machine, and in cloud shell, only to have the same outcome. I tried adding in "Import-Module AzureRm" as an extra line to the runbook immediately prior to the Add-AzAccount line, but this requires user intervention and this script is run in a space which does not allow user intervention. What is required to allow these commands to complete? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7a6394c7-9bef-b8f8-ffd6-9d9d8e2daa07 * Version Independent ID: 5ffa20a2-436c-2726-dc57-9d3b49f9ca39 * Content: [Start an Azure Automation runbook from a webhook](https://docs.microsoft.com/en-us/azure/automation/automation-webhooks) * Content Source: [articles/automation/automation-webhooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-webhooks.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
sample has errors due to powershell commands i am following the steps listed in this document by testing the sample code i have successfully created the webhook runbook and json request in powershell i can verify this as the output in the runbook shows header has required information authenticating to azure with service principal and certificate get connection asset azurerunasconnection authenticating to azure with service principal starting machine name here i have replaced my actual machine name with machine name here for security reasons my input is again i redact the azure webhook storage group and machine name for security reasons uri vms name machine name here resourcegroup storage group name body convertto json inputobject vms header message startedbyuser response invoke webrequest method post uri uri body body headers header jobid convertfrom json response content jobids this shows as completed in the runbook but in the errors section i see this start azaccount the term start azaccount is not recognized as the name of a cmdlet function script file or operable program check the spelling of the name or if a path was included verify that the path is correct and try again at line char start azaccount serviceprincipal tenant conn tenantid categoryinfo objectnotfound start azaccount string commandnotfoundexception fullyqualifiederrorid commandnotfoundexception start azvm the term start azvm is not recognized as the name of a cmdlet function script file or operable program check the spelling of the name or if a path was included verify that the path is correct and try again at line char start azvm name vm name resourcegroup vm resourcegrou categoryinfo objectnotfound start azvm string commandnotfoundexception fullyqualifiederrorid commandnotfoundexception i have tried running this from multiple locations and i receive the same error each time i tried running this from azure cloud shell and received the same outcome i have added az compute before running the task on my local machine and in cloud shell only to have the same outcome i tried adding in import module azurerm as an extra line to the runbook immediately prior to the add azaccount line but this requires user intervention and this script is run in a space which does not allow user intervention what is required to allow these commands to complete document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
52,820
7,786,242,441
IssuesEvent
2018-06-06 18:19:10
golang/go
https://api.github.com/repos/golang/go
closed
database/sql: DB documentation should mention DB.Conn()
Documentation NeedsFix help wanted
https://golang.org/pkg/database/sql/#DB currently states: > If the database has a concept of per-connection state, such state can only be reliably observed within a transaction. This is no longer true: DB.Conn() was added to allow one to have a per-connection state without using transactions. DB documentation should be updated.
1.0
database/sql: DB documentation should mention DB.Conn() - https://golang.org/pkg/database/sql/#DB currently states: > If the database has a concept of per-connection state, such state can only be reliably observed within a transaction. This is no longer true: DB.Conn() was added to allow one to have a per-connection state without using transactions. DB documentation should be updated.
non_process
database sql db documentation should mention db conn currently states if the database has a concept of per connection state such state can only be reliably observed within a transaction this is no longer true db conn was added to allow one to have a per connection state without using transactions db documentation should be updated
0
14,356
17,379,052,384
IssuesEvent
2021-07-31 09:50:22
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Data finder: add test cases for observations, SMHi, BSC, CMIP6
preprocessor
If anyone has access to one of the SMHI/BSC filesystems or CMIP6/OBS/ana4mips/obs4mips/CCMVal1/CCMVal2/EMAC datasets, can you add test cases in the file `tests/integration/data_finder.yml`?
1.0
Data finder: add test cases for observations, SMHi, BSC, CMIP6 - If anyone has access to one of the SMHI/BSC filesystems or CMIP6/OBS/ana4mips/obs4mips/CCMVal1/CCMVal2/EMAC datasets, can you add test cases in the file `tests/integration/data_finder.yml`?
process
data finder add test cases for observations smhi bsc if anyone has access to one of the smhi bsc filesystems or obs emac datasets can you add test cases in the file tests integration data finder yml
1
20,251
11,424,380,394
IssuesEvent
2020-02-03 17:37:42
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
Remove unnecessary OS profile configuration block requirement
breaking-change bug microsoft/3 service/virtual-machine
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The terraform azurerm provider requires a `os_profile_windows_config` configuration block to work, however, the data in this block is completely optional. Therefore an empty block satisfies this. `os_profile_windows_config {}` ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_virtual_machine ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> * https://www.terraform.io/docs/providers/azurerm/r/virtual_machine.html#os_profile_windows_config
1.0
Remove unnecessary OS profile configuration block requirement - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The terraform azurerm provider requires a `os_profile_windows_config` configuration block to work, however, the data in this block is completely optional. Therefore an empty block satisfies this. `os_profile_windows_config {}` ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_virtual_machine ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> * https://www.terraform.io/docs/providers/azurerm/r/virtual_machine.html#os_profile_windows_config
non_process
remove unnecessary os profile configuration block requirement community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the terraform azurerm provider requires a os profile windows config configuration block to work however the data in this block is completely optional therefore an empty block satisfies this os profile windows config new or affected resource s azurerm virtual machine references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
0
15,169
18,939,536,586
IssuesEvent
2021-11-18 00:09:09
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Change in environment creation - no documented workaround
devops/prod product-feedback cba devops-cicd-process/tech
In the referenced documentation > Previously, we handled this case by adding all the project contributors to the administrator role of the environment. Any member of the project could then change these permissions and prevent others from accessing the environment We want the old behavior. Now we have to manually create the environment, and worse, only the creator gets admin access. We want all contributors to have access to it. The documentation does not say how to keep the old behaviour --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Create target environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Change in environment creation - no documented workaround - In the referenced documentation > Previously, we handled this case by adding all the project contributors to the administrator role of the environment. Any member of the project could then change these permissions and prevent others from accessing the environment We want the old behavior. Now we have to manually create the environment, and worse, only the creator gets admin access. We want all contributors to have access to it. The documentation does not say how to keep the old behaviour --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Create target environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
change in environment creation no documented workaround in the referenced documentation previously we handled this case by adding all the project contributors to the administrator role of the environment any member of the project could then change these permissions and prevent others from accessing the environment we want the old behavior now we have to manually create the environment and worse only the creator gets admin access we want all contributors to have access to it the documentation does not say how to keep the old behaviour document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
21,774
30,288,683,685
IssuesEvent
2023-07-09 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Fri, 7 Jul 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### GNEP Based Dynamic Segmentation and Motion Estimation for Neuromorphic Imaging - **Authors:** Harbir Antil, David Sayre - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computer Science and Game Theory (cs.GT); Optimization and Control (math.OC) - **Arxiv link:** https://arxiv.org/abs/2307.02595 - **Pdf link:** https://arxiv.org/pdf/2307.02595 - **Abstract** This paper explores the application of event-based cameras in the domains of image segmentation and motion estimation. These cameras offer a groundbreaking technology by capturing visual information as a continuous stream of asynchronous events, departing from the conventional frame-based image acquisition. We introduce a Generalized Nash Equilibrium based framework that leverages the temporal and spatial information derived from the event stream to carry out segmentation and velocity estimation. To establish the theoretical foundations, we derive an existence criteria and propose a multi-level optimization method for calculating equilibrium. The efficacy of this approach is shown through a series of experiments. ### Zero-Shot Dense Video Captioning by Jointly Optimizing Text and Moment - **Authors:** Yongrae Jo, Seongyun Lee, Aiden SJ Lee, Hyunji Lee, Hanseok Oh, Minjoon Seo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2307.02682 - **Pdf link:** https://arxiv.org/pdf/2307.02682 - **Abstract** Dense video captioning, a task of localizing meaningful moments and generating relevant captions for videos, often requires a large, expensive corpus of annotated video segments paired with text. In an effort to minimize the annotation cost, we propose ZeroTA, a novel method for dense video captioning in a zero-shot manner. Our method does not require any videos or annotations for training; instead, it localizes and describes events within each input video at test time by optimizing solely on the input. This is accomplished by introducing a soft moment mask that represents a temporal segment in the video and jointly optimizing it with the prefix parameters of a language model. This joint optimization aligns a frozen language generation model (i.e., GPT-2) with a frozen vision-language contrastive model (i.e., CLIP) by maximizing the matching score between the generated text and a moment within the video. We also introduce a pairwise temporal IoU loss to let a set of soft moment masks capture multiple distinct events within the video. Our method effectively discovers diverse significant events within the video, with the resulting captions appropriately describing these events. The empirical results demonstrate that ZeroTA surpasses zero-shot baselines and even outperforms the state-of-the-art few-shot method on the widely-used benchmark ActivityNet Captions. Moreover, our method shows greater robustness compared to supervised methods when evaluated in out-of-domain scenarios. This research provides insight into the potential of aligning widely-used models, such as language generation models and vision-language models, to unlock a new capability: understanding temporal aspects of videos. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### MRecGen: Multimodal Appropriate Reaction Generator - **Authors:** Jiaqi Xu, Cheng Luo, Weicheng Xie, Linlin Shen, Xiaofeng Liu, Lu Liu, Hatice Gunes, Siyang Song - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.02609 - **Pdf link:** https://arxiv.org/pdf/2307.02609 - **Abstract** Verbal and non-verbal human reaction generation is a challenging task, as different reactions could be appropriate for responding to the same behaviour. This paper proposes the first multiple and multimodal (verbal and nonverbal) appropriate human reaction generation framework that can generate appropriate and realistic human-style reactions (displayed in the form of synchronised text, audio and video streams) in response to an input user behaviour. This novel technique can be applied to various human-computer interaction scenarios by generating appropriate virtual agent/robot behaviours. Our demo is available at \url{https://github.com/SSYSteve/MRecGen}. ### The Role of Subgroup Separability in Group-Fair Medical Image Classification - **Authors:** Charles Jones, Mélanie Roschewitz, Ben Glocker - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2307.02791 - **Pdf link:** https://arxiv.org/pdf/2307.02791 - **Abstract** We investigate performance disparities in deep classifiers. We find that the ability of classifiers to separate individuals into subgroups varies substantially across medical imaging modalities and protected characteristics; crucially, we show that this property is predictive of algorithmic bias. Through theoretical analysis and extensive empirical evaluation, we find a relationship between subgroup separability, subgroup disparities, and performance degradation when models are trained on data with systematic bias such as underdiagnosis. Our findings shed new light on the question of how models become biased, providing important insights for the development of fair medical imaging AI. ### On the Cultural Gap in Text-to-Image Generation - **Authors:** Bingshuai Liu, Longyue Wang, Chenyang Lyu, Yong Zhang, Jinsong Su, Shuming Shi, Zhaopeng Tu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2307.02971 - **Pdf link:** https://arxiv.org/pdf/2307.02971 - **Abstract** One challenge in text-to-image (T2I) generation is the inadvertent reflection of culture gaps present in the training data, which signifies the disparity in generated image quality when the cultural elements of the input text are rarely collected in the training set. Although various T2I models have shown impressive but arbitrary examples, there is no benchmark to systematically evaluate a T2I model's ability to generate cross-cultural images. To bridge the gap, we propose a Challenging Cross-Cultural (C3) benchmark with comprehensive evaluation criteria, which can assess how well-suited a model is to a target culture. By analyzing the flawed images generated by the Stable Diffusion model on the C3 benchmark, we find that the model often fails to generate certain cultural objects. Accordingly, we propose a novel multi-modal metric that considers object-text alignment to filter the fine-tuning data in the target culture, which is used to fine-tune a T2I model to improve cross-cultural generation. Experimental results show that our multi-modal metric provides stronger data selection performance on the C3 benchmark than existing metrics, in which the object-text alignment is crucial. We release the benchmark, data, code, and generated images to facilitate future research on culturally diverse T2I generation (https://github.com/longyuewangdcu/C3-Bench). ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW There is no result ## Keyword: raw image There is no result
2.0
New submissions for Fri, 7 Jul 23 - ## Keyword: events ### GNEP Based Dynamic Segmentation and Motion Estimation for Neuromorphic Imaging - **Authors:** Harbir Antil, David Sayre - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computer Science and Game Theory (cs.GT); Optimization and Control (math.OC) - **Arxiv link:** https://arxiv.org/abs/2307.02595 - **Pdf link:** https://arxiv.org/pdf/2307.02595 - **Abstract** This paper explores the application of event-based cameras in the domains of image segmentation and motion estimation. These cameras offer a groundbreaking technology by capturing visual information as a continuous stream of asynchronous events, departing from the conventional frame-based image acquisition. We introduce a Generalized Nash Equilibrium based framework that leverages the temporal and spatial information derived from the event stream to carry out segmentation and velocity estimation. To establish the theoretical foundations, we derive an existence criteria and propose a multi-level optimization method for calculating equilibrium. The efficacy of this approach is shown through a series of experiments. ### Zero-Shot Dense Video Captioning by Jointly Optimizing Text and Moment - **Authors:** Yongrae Jo, Seongyun Lee, Aiden SJ Lee, Hyunji Lee, Hanseok Oh, Minjoon Seo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2307.02682 - **Pdf link:** https://arxiv.org/pdf/2307.02682 - **Abstract** Dense video captioning, a task of localizing meaningful moments and generating relevant captions for videos, often requires a large, expensive corpus of annotated video segments paired with text. In an effort to minimize the annotation cost, we propose ZeroTA, a novel method for dense video captioning in a zero-shot manner. Our method does not require any videos or annotations for training; instead, it localizes and describes events within each input video at test time by optimizing solely on the input. This is accomplished by introducing a soft moment mask that represents a temporal segment in the video and jointly optimizing it with the prefix parameters of a language model. This joint optimization aligns a frozen language generation model (i.e., GPT-2) with a frozen vision-language contrastive model (i.e., CLIP) by maximizing the matching score between the generated text and a moment within the video. We also introduce a pairwise temporal IoU loss to let a set of soft moment masks capture multiple distinct events within the video. Our method effectively discovers diverse significant events within the video, with the resulting captions appropriately describing these events. The empirical results demonstrate that ZeroTA surpasses zero-shot baselines and even outperforms the state-of-the-art few-shot method on the widely-used benchmark ActivityNet Captions. Moreover, our method shows greater robustness compared to supervised methods when evaluated in out-of-domain scenarios. This research provides insight into the potential of aligning widely-used models, such as language generation models and vision-language models, to unlock a new capability: understanding temporal aspects of videos. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### MRecGen: Multimodal Appropriate Reaction Generator - **Authors:** Jiaqi Xu, Cheng Luo, Weicheng Xie, Linlin Shen, Xiaofeng Liu, Lu Liu, Hatice Gunes, Siyang Song - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.02609 - **Pdf link:** https://arxiv.org/pdf/2307.02609 - **Abstract** Verbal and non-verbal human reaction generation is a challenging task, as different reactions could be appropriate for responding to the same behaviour. This paper proposes the first multiple and multimodal (verbal and nonverbal) appropriate human reaction generation framework that can generate appropriate and realistic human-style reactions (displayed in the form of synchronised text, audio and video streams) in response to an input user behaviour. This novel technique can be applied to various human-computer interaction scenarios by generating appropriate virtual agent/robot behaviours. Our demo is available at \url{https://github.com/SSYSteve/MRecGen}. ### The Role of Subgroup Separability in Group-Fair Medical Image Classification - **Authors:** Charles Jones, Mélanie Roschewitz, Ben Glocker - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2307.02791 - **Pdf link:** https://arxiv.org/pdf/2307.02791 - **Abstract** We investigate performance disparities in deep classifiers. We find that the ability of classifiers to separate individuals into subgroups varies substantially across medical imaging modalities and protected characteristics; crucially, we show that this property is predictive of algorithmic bias. Through theoretical analysis and extensive empirical evaluation, we find a relationship between subgroup separability, subgroup disparities, and performance degradation when models are trained on data with systematic bias such as underdiagnosis. Our findings shed new light on the question of how models become biased, providing important insights for the development of fair medical imaging AI. ### On the Cultural Gap in Text-to-Image Generation - **Authors:** Bingshuai Liu, Longyue Wang, Chenyang Lyu, Yong Zhang, Jinsong Su, Shuming Shi, Zhaopeng Tu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2307.02971 - **Pdf link:** https://arxiv.org/pdf/2307.02971 - **Abstract** One challenge in text-to-image (T2I) generation is the inadvertent reflection of culture gaps present in the training data, which signifies the disparity in generated image quality when the cultural elements of the input text are rarely collected in the training set. Although various T2I models have shown impressive but arbitrary examples, there is no benchmark to systematically evaluate a T2I model's ability to generate cross-cultural images. To bridge the gap, we propose a Challenging Cross-Cultural (C3) benchmark with comprehensive evaluation criteria, which can assess how well-suited a model is to a target culture. By analyzing the flawed images generated by the Stable Diffusion model on the C3 benchmark, we find that the model often fails to generate certain cultural objects. Accordingly, we propose a novel multi-modal metric that considers object-text alignment to filter the fine-tuning data in the target culture, which is used to fine-tune a T2I model to improve cross-cultural generation. Experimental results show that our multi-modal metric provides stronger data selection performance on the C3 benchmark than existing metrics, in which the object-text alignment is crucial. We release the benchmark, data, code, and generated images to facilitate future research on culturally diverse T2I generation (https://github.com/longyuewangdcu/C3-Bench). ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW There is no result ## Keyword: raw image There is no result
process
new submissions for fri jul keyword events gnep based dynamic segmentation and motion estimation for neuromorphic imaging authors harbir antil david sayre subjects computer vision and pattern recognition cs cv computer science and game theory cs gt optimization and control math oc arxiv link pdf link abstract this paper explores the application of event based cameras in the domains of image segmentation and motion estimation these cameras offer a groundbreaking technology by capturing visual information as a continuous stream of asynchronous events departing from the conventional frame based image acquisition we introduce a generalized nash equilibrium based framework that leverages the temporal and spatial information derived from the event stream to carry out segmentation and velocity estimation to establish the theoretical foundations we derive an existence criteria and propose a multi level optimization method for calculating equilibrium the efficacy of this approach is shown through a series of experiments zero shot dense video captioning by jointly optimizing text and moment authors yongrae jo seongyun lee aiden sj lee hyunji lee hanseok oh minjoon seo subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract dense video captioning a task of localizing meaningful moments and generating relevant captions for videos often requires a large expensive corpus of annotated video segments paired with text in an effort to minimize the annotation cost we propose zerota a novel method for dense video captioning in a zero shot manner our method does not require any videos or annotations for training instead it localizes and describes events within each input video at test time by optimizing solely on the input this is accomplished by introducing a soft moment mask that represents a temporal segment in the video and jointly optimizing it with the prefix parameters of a language model this joint optimization aligns a frozen language generation model i e gpt with a frozen vision language contrastive model i e clip by maximizing the matching score between the generated text and a moment within the video we also introduce a pairwise temporal iou loss to let a set of soft moment masks capture multiple distinct events within the video our method effectively discovers diverse significant events within the video with the resulting captions appropriately describing these events the empirical results demonstrate that zerota surpasses zero shot baselines and even outperforms the state of the art few shot method on the widely used benchmark activitynet captions moreover our method shows greater robustness compared to supervised methods when evaluated in out of domain scenarios this research provides insight into the potential of aligning widely used models such as language generation models and vision language models to unlock a new capability understanding temporal aspects of videos keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp mrecgen multimodal appropriate reaction generator authors jiaqi xu cheng luo weicheng xie linlin shen xiaofeng liu lu liu hatice gunes siyang song subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract verbal and non verbal human reaction generation is a challenging task as different reactions could be appropriate for responding to the same behaviour this paper proposes the first multiple and multimodal verbal and nonverbal appropriate human reaction generation framework that can generate appropriate and realistic human style reactions displayed in the form of synchronised text audio and video streams in response to an input user behaviour this novel technique can be applied to various human computer interaction scenarios by generating appropriate virtual agent robot behaviours our demo is available at url the role of subgroup separability in group fair medical image classification authors charles jones mélanie roschewitz ben glocker subjects computer vision and pattern recognition cs cv artificial intelligence cs ai computers and society cs cy machine learning cs lg arxiv link pdf link abstract we investigate performance disparities in deep classifiers we find that the ability of classifiers to separate individuals into subgroups varies substantially across medical imaging modalities and protected characteristics crucially we show that this property is predictive of algorithmic bias through theoretical analysis and extensive empirical evaluation we find a relationship between subgroup separability subgroup disparities and performance degradation when models are trained on data with systematic bias such as underdiagnosis our findings shed new light on the question of how models become biased providing important insights for the development of fair medical imaging ai on the cultural gap in text to image generation authors bingshuai liu longyue wang chenyang lyu yong zhang jinsong su shuming shi zhaopeng tu subjects computer vision and pattern recognition cs cv artificial intelligence cs ai computation and language cs cl arxiv link pdf link abstract one challenge in text to image generation is the inadvertent reflection of culture gaps present in the training data which signifies the disparity in generated image quality when the cultural elements of the input text are rarely collected in the training set although various models have shown impressive but arbitrary examples there is no benchmark to systematically evaluate a model s ability to generate cross cultural images to bridge the gap we propose a challenging cross cultural benchmark with comprehensive evaluation criteria which can assess how well suited a model is to a target culture by analyzing the flawed images generated by the stable diffusion model on the benchmark we find that the model often fails to generate certain cultural objects accordingly we propose a novel multi modal metric that considers object text alignment to filter the fine tuning data in the target culture which is used to fine tune a model to improve cross cultural generation experimental results show that our multi modal metric provides stronger data selection performance on the benchmark than existing metrics in which the object text alignment is crucial we release the benchmark data code and generated images to facilitate future research on culturally diverse generation keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw there is no result keyword raw image there is no result
1
9,659
11,726,051,416
IssuesEvent
2020-03-10 13:57:13
oracle/truffleruby
https://api.github.com/repos/oracle/truffleruby
closed
Frozen string literal modification error in place MRI doesn't
compatibility
https://travis-ci.org/puma/puma/jobs/659981141#L432 We run all our CI jobs with --enable-frozen-string-literal. This [doesnt blow up](https://github.com/puma/puma/runs/494216825#step:6:16) when running Rubocop for us on MRI.
True
Frozen string literal modification error in place MRI doesn't - https://travis-ci.org/puma/puma/jobs/659981141#L432 We run all our CI jobs with --enable-frozen-string-literal. This [doesnt blow up](https://github.com/puma/puma/runs/494216825#step:6:16) when running Rubocop for us on MRI.
non_process
frozen string literal modification error in place mri doesn t we run all our ci jobs with enable frozen string literal this when running rubocop for us on mri
0
18,398
24,534,337,476
IssuesEvent
2022-10-11 19:16:18
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
About the split expression
doc-enhancement devops/prod Pri1 devops-cicd-process/tech
The documentation does not clarify whether it is possible to use the split expression in any other scenario other than in each keyword, I would like for example to be able to split an Azure resource-id and then fetch index 8 which would be the resource-name, would that be possible? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6 * Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18 * Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops&source=docs#split) * Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/expressions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
About the split expression - The documentation does not clarify whether it is possible to use the split expression in any other scenario other than in each keyword, I would like for example to be able to split an Azure resource-id and then fetch index 8 which would be the resource-name, would that be possible? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6 * Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18 * Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops&source=docs#split) * Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/expressions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
about the split expression the documentation does not clarify whether it is possible to use the split expression in any other scenario other than in each keyword i would like for example to be able to split an azure resource id and then fetch index which would be the resource name would that be possible document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
298,536
25,835,501,512
IssuesEvent
2022-12-12 19:17:47
tricia-holmes/rancid-tomatillos
https://api.github.com/repos/tricia-holmes/rancid-tomatillos
opened
Need to add tests for Movie Details Content and SearchBar
testing
- Movie Details now as the additional content - Search bar feature has now been added to the Home page
1.0
Need to add tests for Movie Details Content and SearchBar - - Movie Details now as the additional content - Search bar feature has now been added to the Home page
non_process
need to add tests for movie details content and searchbar movie details now as the additional content search bar feature has now been added to the home page
0
55,044
14,146,441,427
IssuesEvent
2020-11-10 19:13:36
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Generate behavior changes from 3.13.3 to 3.14.2
T: Defect
### Expected behavior Records should be generated extending `UpdatableRecordImpl`. (they are generated like this in 3.13.3) ### Actual behavior Records only extend `TableRecordImpl`, although a Primary key is in the table and the config hasn't changed. ### Versions - jOOQ: 3.14.2 - Java: openjdk 13.0.2 2020-01-14 - Database (include vendor): PostgreSQL 12 - OS: MacOS - JDBC Driver (include name if inofficial driver): org.postgresql:postgresql:42.2.14
1.0
Generate behavior changes from 3.13.3 to 3.14.2 - ### Expected behavior Records should be generated extending `UpdatableRecordImpl`. (they are generated like this in 3.13.3) ### Actual behavior Records only extend `TableRecordImpl`, although a Primary key is in the table and the config hasn't changed. ### Versions - jOOQ: 3.14.2 - Java: openjdk 13.0.2 2020-01-14 - Database (include vendor): PostgreSQL 12 - OS: MacOS - JDBC Driver (include name if inofficial driver): org.postgresql:postgresql:42.2.14
non_process
generate behavior changes from to expected behavior records should be generated extending updatablerecordimpl they are generated like this in actual behavior records only extend tablerecordimpl although a primary key is in the table and the config hasn t changed versions jooq java openjdk database include vendor postgresql os macos jdbc driver include name if inofficial driver org postgresql postgresql
0
39,145
5,040,451,208
IssuesEvent
2016-12-19 05:38:36
hypermodules/hyperamp
https://api.github.com/repos/hypermodules/hyperamp
closed
Sticky table headers
design enhancement
We need to do sticky table headers, so you can adjust the sort column and direction when you are at any scroll position. Ideally, we could just use something like [Sticky Positioning](https://developer.mozilla.org/en-US/docs/Web/CSS/position#Sticky_positioning), but it doesn't work on `thead` elements, and it doesn't work in chrome webkit! 😭 There are a bunch of [hacks](http://stackoverflow.com/questions/1030043/html-table-headers-always-visible-at-top-of-window-when-viewing-a-large-table) that you can do to get it to sort of work with the table elements, but they are all pretty gross and don't perform great. The third options I was thinking of is dropping the `table` component all together and using flex'd div's with `overflow: auto` set on the table row container. We are going to have to port over/redo table styles, but will ultimately set us to up to do fun/complicated table rows without having to fight the limitations of the archaic table element. e.g. <img width="1139" alt="screenshot 2016-08-14 21 48 23" src="https://cloud.githubusercontent.com/assets/166301/17655925/dc835b64-6268-11e6-840c-4c9e8a25edd5.png">
1.0
Sticky table headers - We need to do sticky table headers, so you can adjust the sort column and direction when you are at any scroll position. Ideally, we could just use something like [Sticky Positioning](https://developer.mozilla.org/en-US/docs/Web/CSS/position#Sticky_positioning), but it doesn't work on `thead` elements, and it doesn't work in chrome webkit! 😭 There are a bunch of [hacks](http://stackoverflow.com/questions/1030043/html-table-headers-always-visible-at-top-of-window-when-viewing-a-large-table) that you can do to get it to sort of work with the table elements, but they are all pretty gross and don't perform great. The third options I was thinking of is dropping the `table` component all together and using flex'd div's with `overflow: auto` set on the table row container. We are going to have to port over/redo table styles, but will ultimately set us to up to do fun/complicated table rows without having to fight the limitations of the archaic table element. e.g. <img width="1139" alt="screenshot 2016-08-14 21 48 23" src="https://cloud.githubusercontent.com/assets/166301/17655925/dc835b64-6268-11e6-840c-4c9e8a25edd5.png">
non_process
sticky table headers we need to do sticky table headers so you can adjust the sort column and direction when you are at any scroll position ideally we could just use something like but it doesn t work on thead elements and it doesn t work in chrome webkit 😭 there are a bunch of that you can do to get it to sort of work with the table elements but they are all pretty gross and don t perform great the third options i was thinking of is dropping the table component all together and using flex d div s with overflow auto set on the table row container we are going to have to port over redo table styles but will ultimately set us to up to do fun complicated table rows without having to fight the limitations of the archaic table element e g img width alt screenshot src
0
157,939
19,988,935,152
IssuesEvent
2022-01-31 02:05:41
sureng-ws-ibm/sg-x-nodegoat
https://api.github.com/repos/sureng-ws-ibm/sg-x-nodegoat
closed
CVE-2019-20922 (High) detected in handlebars-4.0.5.tgz - autoclosed
security vulnerability
## CVE-2019-20922 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/nyc/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - istanbul-reports-1.0.0-alpha.8.tgz - :x: **handlebars-4.0.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sureng-ws-ibm/sg-x-nodegoat/commit/81cd799c64e09f7670559bce715f20a9ec7821d5">81cd799c64e09f7670559bce715f20a9ec7821d5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 4.4.5 allows Regular Expression Denial of Service (ReDoS) because of eager matching. The parser may be forced into an endless loop while processing crafted templates. This may allow attackers to exhaust system resources. <p>Publish Date: 2020-09-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20922>CVE-2019-20922</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: handlebars - 4.4.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-reports:1.0.0-alpha.8;handlebars:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.4.5","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20922","vulnerabilityDetails":"Handlebars before 4.4.5 allows Regular Expression Denial of Service (ReDoS) because of eager matching. The parser may be forced into an endless loop while processing crafted templates. This may allow attackers to exhaust system resources.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20922","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-20922 (High) detected in handlebars-4.0.5.tgz - autoclosed - ## CVE-2019-20922 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/nyc/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - istanbul-reports-1.0.0-alpha.8.tgz - :x: **handlebars-4.0.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sureng-ws-ibm/sg-x-nodegoat/commit/81cd799c64e09f7670559bce715f20a9ec7821d5">81cd799c64e09f7670559bce715f20a9ec7821d5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 4.4.5 allows Regular Expression Denial of Service (ReDoS) because of eager matching. The parser may be forced into an endless loop while processing crafted templates. This may allow attackers to exhaust system resources. <p>Publish Date: 2020-09-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20922>CVE-2019-20922</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: handlebars - 4.4.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;istanbul-reports:1.0.0-alpha.8;handlebars:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.4.5","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20922","vulnerabilityDetails":"Handlebars before 4.4.5 allows Regular Expression Denial of Service (ReDoS) because of eager matching. The parser may be forced into an endless loop while processing crafted templates. This may allow attackers to exhaust system resources.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20922","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in handlebars tgz autoclosed cve high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file package json path to vulnerable library node modules nyc node modules handlebars package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz istanbul reports alpha tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details handlebars before allows regular expression denial of service redos because of eager matching the parser may be forced into an endless loop while processing crafted templates this may allow attackers to exhaust system resources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt if grunt contrib nodeunit nodeunit tap nyc istanbul reports alpha handlebars isminimumfixversionavailable true minimumfixversion handlebars isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails handlebars before allows regular expression denial of service redos because of eager matching the parser may be forced into an endless loop while processing crafted templates this may allow attackers to exhaust system resources vulnerabilityurl
0
18,683
24,594,848,734
IssuesEvent
2022-10-14 07:24:09
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Android] [FHIR] Questionnaire response > Display text is getting displayed when participant select 'Other' option in the text choice questionnaire
Bug P1 Android Response datastore Process: Fixed Process: Tested QA Process: Tested dev
**Pre-condition:** Text choice multiselect with other option questionnaire should be configured in the Study builder **Steps:** 1. Sign up or Sign to the mobile app 2. Enroll to the study 3. Submit the response for created questionnaire 4. Go to the FHIR store 5. Go to questionnaire response 6. Click on the particular record and observe the JSON file **AR:** Questionnaire response > Display text which is configured in the SB is getting displayed when participant select 'Other' option in the text choice questionnaire **ER:** Questionnaire response > Value which is configured in the SB should get displayed when participant select 'Other' option in the text choice questionnaire ![A](https://user-images.githubusercontent.com/86007179/191544385-fcc07e10-9c60-493f-836d-b7cad6ffabc5.png)
3.0
[Android] [FHIR] Questionnaire response > Display text is getting displayed when participant select 'Other' option in the text choice questionnaire - **Pre-condition:** Text choice multiselect with other option questionnaire should be configured in the Study builder **Steps:** 1. Sign up or Sign to the mobile app 2. Enroll to the study 3. Submit the response for created questionnaire 4. Go to the FHIR store 5. Go to questionnaire response 6. Click on the particular record and observe the JSON file **AR:** Questionnaire response > Display text which is configured in the SB is getting displayed when participant select 'Other' option in the text choice questionnaire **ER:** Questionnaire response > Value which is configured in the SB should get displayed when participant select 'Other' option in the text choice questionnaire ![A](https://user-images.githubusercontent.com/86007179/191544385-fcc07e10-9c60-493f-836d-b7cad6ffabc5.png)
process
questionnaire response display text is getting displayed when participant select other option in the text choice questionnaire pre condition text choice multiselect with other option questionnaire should be configured in the study builder steps sign up or sign to the mobile app enroll to the study submit the response for created questionnaire go to the fhir store go to questionnaire response click on the particular record and observe the json file ar questionnaire response display text which is configured in the sb is getting displayed when participant select other option in the text choice questionnaire er questionnaire response value which is configured in the sb should get displayed when participant select other option in the text choice questionnaire
1
82,258
10,237,203,938
IssuesEvent
2019-08-19 13:26:31
LSCOIT/LegalNavigator
https://api.github.com/repos/LSCOIT/LegalNavigator
closed
As Portal User, I want to see a message displayed in the portal when I entered a location so that I would know if it is supported or not supported.
Backlog Design Change PILOT
_From @czdiana2321 on May 21, 2018 19:39_ USER->UI Acceptance Criteria: 1. The system is able to display content that is relevant globally. 2. The system is able to display a message if the location entered is supported or not. 3. The Portal User is able to enter location to display content that is relevant globally. 4. The Portal User is able to view a message if the location entered is supported or not supported. _Copied from original issue: Microsoft/Access2Justice#82_
1.0
As Portal User, I want to see a message displayed in the portal when I entered a location so that I would know if it is supported or not supported. - _From @czdiana2321 on May 21, 2018 19:39_ USER->UI Acceptance Criteria: 1. The system is able to display content that is relevant globally. 2. The system is able to display a message if the location entered is supported or not. 3. The Portal User is able to enter location to display content that is relevant globally. 4. The Portal User is able to view a message if the location entered is supported or not supported. _Copied from original issue: Microsoft/Access2Justice#82_
non_process
as portal user i want to see a message displayed in the portal when i entered a location so that i would know if it is supported or not supported from on may user ui acceptance criteria the system is able to display content that is relevant globally the system is able to display a message if the location entered is supported or not the portal user is able to enter location to display content that is relevant globally the portal user is able to view a message if the location entered is supported or not supported copied from original issue microsoft
0
355,576
10,582,500,196
IssuesEvent
2019-10-08 11:42:06
fack2/attention-training
https://api.github.com/repos/fack2/attention-training
closed
BUG: update travis code to test node and react packages.
bug done priority-1
Our latest change requires us to update the travis CI configs. ## How it works: We add an ```env``` key in the travis config to do two different builds * client directoy * server directory These two directories will instruct travis to run two different jobs -one for the client and one for the server In the script key, we give travis instruction of what to do in each build. * since we don't have any tests yet, we will run the ```install``` and ```build``` for react, and only the ```install``` script for nodejs ## The travis file should look like this ``` language: node_js node_js: - "node" env: - BUILD_DIR=client - BUILD_DIR=server script: - if [ "$BUILD_DIR" = "client" ]; then cd $BUILD_DIR && npm install && npm run build; fi - if [ "$BUILD_DIR" = "server" ]; then cd $BUILD_DIR && npm install; fi ``` ### Don't copy anything that isn't clear to you. Ask me anything.
1.0
BUG: update travis code to test node and react packages. - Our latest change requires us to update the travis CI configs. ## How it works: We add an ```env``` key in the travis config to do two different builds * client directoy * server directory These two directories will instruct travis to run two different jobs -one for the client and one for the server In the script key, we give travis instruction of what to do in each build. * since we don't have any tests yet, we will run the ```install``` and ```build``` for react, and only the ```install``` script for nodejs ## The travis file should look like this ``` language: node_js node_js: - "node" env: - BUILD_DIR=client - BUILD_DIR=server script: - if [ "$BUILD_DIR" = "client" ]; then cd $BUILD_DIR && npm install && npm run build; fi - if [ "$BUILD_DIR" = "server" ]; then cd $BUILD_DIR && npm install; fi ``` ### Don't copy anything that isn't clear to you. Ask me anything.
non_process
bug update travis code to test node and react packages our latest change requires us to update the travis ci configs how it works we add an env key in the travis config to do two different builds client directoy server directory these two directories will instruct travis to run two different jobs one for the client and one for the server in the script key we give travis instruction of what to do in each build since we don t have any tests yet we will run the install and build for react and only the install script for nodejs the travis file should look like this language node js node js node env build dir client build dir server script if then cd build dir npm install npm run build fi if then cd build dir npm install fi don t copy anything that isn t clear to you ask me anything
0
22,616
31,842,560,988
IssuesEvent
2023-09-14 17:23:05
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
jfrog-cli-v2 2.48.0 has 4 guarddog issues
npm-install-script npm-silent-process-execution exec-base64
```{"exec-base64":[{"code":" retcode = subprocess.call(command, cwd=ROOT_DIR, env=env, shell=True)","location":"package/node-v8.11.1-linux-x64/lib/node_modules/npm/node_modules/node-gyp/gyp/buildbot/buildbot_run.py:89","message":"This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"}],"npm-install-script":[{"code":" \"prepare\": \"node bin/npm-cli.js --no-timing prune --prefix=. --no-global \u0026\u0026 rimraf test/*/*/node_modules \u0026\u0026 make -j4 doc\",","location":"package/node-v8.11.1-linux-x64/lib/node_modules/npm/package.json:255","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"preinstall\": \"node init.js\"","location":"package/package.json:9","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":"\t\tspawn(process.execPath, [path.join(__dirname, 'check.js'), JSON.stringify(this.options)], {\n\t\t\tdetached: true,\n\t\t\tstdio: 'ignore'\n\t\t}).unref();","location":"package/node-v8.11.1-linux-x64/lib/node_modules/npm/node_modules/update-notifier/index.js:93","message":"This package is silently executing another executable"}]}```
1.0
jfrog-cli-v2 2.48.0 has 4 guarddog issues - ```{"exec-base64":[{"code":" retcode = subprocess.call(command, cwd=ROOT_DIR, env=env, shell=True)","location":"package/node-v8.11.1-linux-x64/lib/node_modules/npm/node_modules/node-gyp/gyp/buildbot/buildbot_run.py:89","message":"This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"}],"npm-install-script":[{"code":" \"prepare\": \"node bin/npm-cli.js --no-timing prune --prefix=. --no-global \u0026\u0026 rimraf test/*/*/node_modules \u0026\u0026 make -j4 doc\",","location":"package/node-v8.11.1-linux-x64/lib/node_modules/npm/package.json:255","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"preinstall\": \"node init.js\"","location":"package/package.json:9","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":"\t\tspawn(process.execPath, [path.join(__dirname, 'check.js'), JSON.stringify(this.options)], {\n\t\t\tdetached: true,\n\t\t\tstdio: 'ignore'\n\t\t}).unref();","location":"package/node-v8.11.1-linux-x64/lib/node_modules/npm/node_modules/update-notifier/index.js:93","message":"This package is silently executing another executable"}]}```
process
jfrog cli has guarddog issues exec npm install script npm silent process execution n t t tdetached true n t t tstdio ignore n t t unref location package node linux lib node modules npm node modules update notifier index js message this package is silently executing another executable
1
10,211
13,068,324,975
IssuesEvent
2020-07-31 03:13:35
nearprotocol/near-web3-provider
https://api.github.com/repos/nearprotocol/near-web3-provider
closed
Find solution to transfer funds from evm_account to near account.
in process
This web3 provider hasn't yet fulfilled ALL the near_evm transfer options. Right now, we have the ability to: * transfer funds to your evm address. * transfer funds between evm addresses But we cannot: * transfer evm address funds back to near account * transfer funds from evm account to corresponding evm account of near account (e.g. input recipient near account in order to send funds to corresponding evm account)
1.0
Find solution to transfer funds from evm_account to near account. - This web3 provider hasn't yet fulfilled ALL the near_evm transfer options. Right now, we have the ability to: * transfer funds to your evm address. * transfer funds between evm addresses But we cannot: * transfer evm address funds back to near account * transfer funds from evm account to corresponding evm account of near account (e.g. input recipient near account in order to send funds to corresponding evm account)
process
find solution to transfer funds from evm account to near account this provider hasn t yet fulfilled all the near evm transfer options right now we have the ability to transfer funds to your evm address transfer funds between evm addresses but we cannot transfer evm address funds back to near account transfer funds from evm account to corresponding evm account of near account e g input recipient near account in order to send funds to corresponding evm account
1
2,275
5,104,271,670
IssuesEvent
2017-01-05 00:27:32
Jarvvski/CavTools
https://api.github.com/repos/Jarvvski/CavTools
closed
Recruit Tracker
Process Flow Request
<h2>Problem</h2><br />Can not tell how many recruits a recruiter has gotten past the current month<br /><hr><h2>Reason</h2><br />Need to award medals for recruiting and also for retention purposes in the department.<br><br>-First Lieutenant Lombardi.M
1.0
Recruit Tracker - <h2>Problem</h2><br />Can not tell how many recruits a recruiter has gotten past the current month<br /><hr><h2>Reason</h2><br />Need to award medals for recruiting and also for retention purposes in the department.<br><br>-First Lieutenant Lombardi.M
process
recruit tracker problem can not tell how many recruits a recruiter has gotten past the current month reason need to award medals for recruiting and also for retention purposes in the department first lieutenant lombardi m
1
253,645
19,144,146,324
IssuesEvent
2021-12-02 04:45:25
IBM-Cloud/terraform-provider-ibm
https://api.github.com/repos/IBM-Cloud/terraform-provider-ibm
closed
Bug in documentation for ibm_access_group_policy
documentation IAMPAP
<!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform IBM Provider Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * ibm_iam_access_group_policy * https://registry.terraform.io/providers/IBM-Cloud/ibm/latest/docs/resources/iam_access_group_policy ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation. ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please share a link to the ZIP file. ``` ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html). ---> ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior <!--- What should have happened? ---> The resource name should be `ibm_iam_access_group_policy` in the resource documentation https://registry.terraform.io/providers/IBM-Cloud/ibm/latest/docs/resources/iam_access_group_policy ### Actual Behavior <!--- What actually happened? ---> The resource name in the title has a typo ibm_access_group_policy. It is missing `iam` keyword ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: ---> * #0000
1.0
Bug in documentation for ibm_access_group_policy - <!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform IBM Provider Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * ibm_iam_access_group_policy * https://registry.terraform.io/providers/IBM-Cloud/ibm/latest/docs/resources/iam_access_group_policy ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation. ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please share a link to the ZIP file. ``` ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html). ---> ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior <!--- What should have happened? ---> The resource name should be `ibm_iam_access_group_policy` in the resource documentation https://registry.terraform.io/providers/IBM-Cloud/ibm/latest/docs/resources/iam_access_group_policy ### Actual Behavior <!--- What actually happened? ---> The resource name in the title has a typo ibm_access_group_policy. It is missing `iam` keyword ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: ---> * #0000
non_process
bug in documentation for ibm access group policy please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform ibm provider version affected resource s ibm iam access group policy terraform configuration files please include all terraform configurations required to reproduce the bug bug reports without a functional reproduction may be closed without investigation hcl copy paste your terraform configurations here for large terraform configs please share a link to the zip file debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the panic output expected behavior the resource name should be ibm iam access group policy in the resource documentation actual behavior the resource name in the title has a typo ibm access group policy it is missing iam keyword steps to reproduce terraform apply important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example
0
140
2,575,854,134
IssuesEvent
2015-02-12 03:09:21
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Failure System.Diagnostics.ProcessTests on Chinese Windows 8.1
grabbed by community System.Diagnostics.Process test bug
Looks like it's because there's localized string for "at" in the exception messages. - System.Diagnostics.ProcessTests.ProcessTest.Process_SyncErrorStream [FAIL] Assert.Contains() Failure Not found: Unhandled Exception: System.Exception: Intentional Exception thrown at ProcessTest_ConsoleApp.Program.Main(String[] args) In value: Unhandled Exception: System.Exception: Intentional Exception thrown ? ProcessTest_ConsoleApp.Program.Main(String[] args) ?? C:\github\j eremymeng\corefx\src\System.Diagnostics.Process\tests\ProcessTest_ConsoleApp\ ProcessTest_ConsoleApp.cs:?? 25 - System.Diagnostics.ProcessTests.ProcessTest.Process_AsyncErrorStream [FAIL ] Assert.Contains() Failure Not found: Unhandled Exception: System.Exception: Intentional Exception thrown at ProcessTest_ConsoleApp.Program.Main(String[] args) In value: Unhandled Exception: System.Exception: Intentional Exception thrown ? ProcessTest_ConsoleApp.Program.Main(String[] args) ?? C:\github\j eremymeng\corefx\src\System.Diagnostics.Process\tests\ProcessTest_ConsoleApp\ ProcessTest_ConsoleApp.cs:?? 25
1.0
Failure System.Diagnostics.ProcessTests on Chinese Windows 8.1 - Looks like it's because there's localized string for "at" in the exception messages. - System.Diagnostics.ProcessTests.ProcessTest.Process_SyncErrorStream [FAIL] Assert.Contains() Failure Not found: Unhandled Exception: System.Exception: Intentional Exception thrown at ProcessTest_ConsoleApp.Program.Main(String[] args) In value: Unhandled Exception: System.Exception: Intentional Exception thrown ? ProcessTest_ConsoleApp.Program.Main(String[] args) ?? C:\github\j eremymeng\corefx\src\System.Diagnostics.Process\tests\ProcessTest_ConsoleApp\ ProcessTest_ConsoleApp.cs:?? 25 - System.Diagnostics.ProcessTests.ProcessTest.Process_AsyncErrorStream [FAIL ] Assert.Contains() Failure Not found: Unhandled Exception: System.Exception: Intentional Exception thrown at ProcessTest_ConsoleApp.Program.Main(String[] args) In value: Unhandled Exception: System.Exception: Intentional Exception thrown ? ProcessTest_ConsoleApp.Program.Main(String[] args) ?? C:\github\j eremymeng\corefx\src\System.Diagnostics.Process\tests\ProcessTest_ConsoleApp\ ProcessTest_ConsoleApp.cs:?? 25
process
failure system diagnostics processtests on chinese windows looks like it s because there s localized string for at in the exception messages system diagnostics processtests processtest process syncerrorstream assert contains failure not found unhandled exception system exception intentional exception thrown at processtest consoleapp program main string args in value unhandled exception system exception intentional exception thrown processtest consoleapp program main string args c github j eremymeng corefx src system diagnostics process tests processtest consoleapp processtest consoleapp cs system diagnostics processtests processtest process asyncerrorstream fail assert contains failure not found unhandled exception system exception intentional exception thrown at processtest consoleapp program main string args in value unhandled exception system exception intentional exception thrown processtest consoleapp program main string args c github j eremymeng corefx src system diagnostics process tests processtest consoleapp processtest consoleapp cs
1
3,896
6,821,436,901
IssuesEvent
2017-11-07 16:45:57
ontop/ontop
https://api.github.com/repos/ontop/ontop
opened
Explicit exception message for unsupported datatypes
status: accepted topic: mapping processing type: enhancement w: error message
When an unsupported RDF datatype is used in a mapping assertion, it currently does not explicitly state the problem: org.semanticweb.owlapi.reasoner.IllegalConfigurationException: it.unibz.inf.ontop.exception.InvalidMappingExceptionWithIndicator: The syntax of the mapping is invalid (and therefore cannot be processed). [List of involved mappings] A clearer (high-level) exception could be thrown.
1.0
Explicit exception message for unsupported datatypes - When an unsupported RDF datatype is used in a mapping assertion, it currently does not explicitly state the problem: org.semanticweb.owlapi.reasoner.IllegalConfigurationException: it.unibz.inf.ontop.exception.InvalidMappingExceptionWithIndicator: The syntax of the mapping is invalid (and therefore cannot be processed). [List of involved mappings] A clearer (high-level) exception could be thrown.
process
explicit exception message for unsupported datatypes when an unsupported rdf datatype is used in a mapping assertion it currently does not explicitly state the problem org semanticweb owlapi reasoner illegalconfigurationexception it unibz inf ontop exception invalidmappingexceptionwithindicator the syntax of the mapping is invalid and therefore cannot be processed a clearer high level exception could be thrown
1
234,825
19,270,482,067
IssuesEvent
2021-12-10 04:16:55
cse110-fa21-group2/cse110-fa21-group2
https://api.github.com/repos/cse110-fa21-group2/cse110-fa21-group2
closed
Britney's Testing Assignments
testing
#Functions to Test - createList(listName) - removeRecipeFromList(listName, recipeId)
1.0
Britney's Testing Assignments - #Functions to Test - createList(listName) - removeRecipeFromList(listName, recipeId)
non_process
britney s testing assignments functions to test createlist listname removerecipefromlist listname recipeid
0
131,306
18,268,599,802
IssuesEvent
2021-10-04 11:24:46
allinbits/cosmos-cash
https://api.github.com/repos/allinbits/cosmos-cash
closed
Ensure that there is at least one verification method for a did document
Design module:did
At the moment (v1.0.0) all the verification methods of a did can be revoked, the facto locking a did document. Shall a did document have at least one verification method OR one controller?
1.0
Ensure that there is at least one verification method for a did document - At the moment (v1.0.0) all the verification methods of a did can be revoked, the facto locking a did document. Shall a did document have at least one verification method OR one controller?
non_process
ensure that there is at least one verification method for a did document at the moment all the verification methods of a did can be revoked the facto locking a did document shall a did document have at least one verification method or one controller
0
86,821
10,518,759,574
IssuesEvent
2019-09-29 13:17:58
lejard-h/chopper
https://api.github.com/repos/lejard-h/chopper
closed
how to increase timeout in chopper
documentation enhancement help wanted no-issue-activity
help or example needed : how to increase timeout in chopper ?
1.0
how to increase timeout in chopper - help or example needed : how to increase timeout in chopper ?
non_process
how to increase timeout in chopper help or example needed how to increase timeout in chopper
0
19,831
26,222,674,837
IssuesEvent
2023-01-04 16:00:20
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
Fix mypy continuous integration issues
process + tools
From https://github.com/pystatgen/sgkit/pull/32: - [x] [mypy precommit](https://github.com/pre-commit/mirrors-mypy) should use the settings in `setup.cfg` - [ ] GH actions should publish the mypy HTML report (cf. [Publishing Test Results](https://github.community/t/publishing-test-results/16215) on the GitHub community Discourse). - [ ] precommit mypy hook requires duplication of requirements
1.0
Fix mypy continuous integration issues - From https://github.com/pystatgen/sgkit/pull/32: - [x] [mypy precommit](https://github.com/pre-commit/mirrors-mypy) should use the settings in `setup.cfg` - [ ] GH actions should publish the mypy HTML report (cf. [Publishing Test Results](https://github.community/t/publishing-test-results/16215) on the GitHub community Discourse). - [ ] precommit mypy hook requires duplication of requirements
process
fix mypy continuous integration issues from should use the settings in setup cfg gh actions should publish the mypy html report cf on the github community discourse precommit mypy hook requires duplication of requirements
1
271,427
29,498,538,225
IssuesEvent
2023-06-02 19:15:59
stefanfreitag/s3_yum_repository
https://api.github.com/repos/stefanfreitag/s3_yum_repository
closed
CVE-2020-7598 (High) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz
security vulnerability
## CVE-2020-7598 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary> <p> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/s3_yum_repository/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/s3_yum_repository/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - ts-jest-25.2.1.tgz (Root Library) - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/s3_yum_repository/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/s3_yum_repository/node_modules/sane/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - ts-jest-25.2.1.tgz (Root Library) - json5-2.1.1.tgz - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/s3_yum_repository/commit/e347c5dd48feb5ebd3ff087bfa625e34b0787230">e347c5dd48feb5ebd3ff087bfa625e34b0787230</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution: minimist - 0.2.1,1.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7598 (High) detected in minimist-0.0.8.tgz, minimist-1.2.0.tgz - ## CVE-2020-7598 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.0.tgz</b></p></summary> <p> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/s3_yum_repository/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/s3_yum_repository/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - ts-jest-25.2.1.tgz (Root Library) - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/s3_yum_repository/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/s3_yum_repository/node_modules/sane/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - ts-jest-25.2.1.tgz (Root Library) - json5-2.1.1.tgz - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/s3_yum_repository/commit/e347c5dd48feb5ebd3ff087bfa625e34b0787230">e347c5dd48feb5ebd3ff087bfa625e34b0787230</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution: minimist - 0.2.1,1.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in minimist tgz minimist tgz cve high severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file tmp ws scm yum repository package json path to vulnerable library tmp ws scm yum repository node modules minimist package json dependency hierarchy ts jest tgz root library mkdirp tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file tmp ws scm yum repository package json path to vulnerable library tmp ws scm yum repository node modules sane node modules minimist package json dependency hierarchy ts jest tgz root library tgz x minimist tgz vulnerable library found in head commit a href vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution minimist step up your open source security game with whitesource
0
152,466
23,979,927,069
IssuesEvent
2022-09-13 14:21:54
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
v2 PUT endpoint for editing text files
area/language solution/studio/designer solution/app-backend team/studio
### Background Today text files are updated by building a new text file in the old text format with the filename format: `resource.*.json`. The endpoint for saving the text is now defined in TextController as a POST request to `SaveResource()` which checks the jsonData from requestbody for duplicate keys and sorts the dictionary before storing it. The operations for storing/overwriting the text exists in a method defined in IRepository, which uses a method in ServiceRepositorySettings for getting the correct path. ### Description The new PUT endpoint for editing a text file in a specific language for an App repository should be created in the new TextsController which handles all interactions with text files in the new format, meaning that it's reached from the version 2 reoute URL: `designer/api/v2/{org}/{repo}/texts`. The endpoint will take a json object consisting of all the ID's and text pairs as key:value pairs as its payload. The json object should overwrite the existing text file for the language defined in the URL route. _NB: new endpoint will not sort the dictionary before storing it, if this is a desired feature it has to be done in frontend_. Endpoint is called repeatedly by some client-triggered action ( e.g. when user stops editing for some while). This includes when editing a value connected to an already existing key/ID, editing an existing key or adding a new key:value pair. All the key:value pairs will be overwritten with each PUT call to the endpoint. ### Additional Information See issue #8808 for an explanation of the conversion from old text format to new text format. ### Tasks _No response_ ### Acceptance Criterias - [ ] Implementation of EditText() endpoint - [ ] Unit/integration tests for the endpoint
1.0
v2 PUT endpoint for editing text files - ### Background Today text files are updated by building a new text file in the old text format with the filename format: `resource.*.json`. The endpoint for saving the text is now defined in TextController as a POST request to `SaveResource()` which checks the jsonData from requestbody for duplicate keys and sorts the dictionary before storing it. The operations for storing/overwriting the text exists in a method defined in IRepository, which uses a method in ServiceRepositorySettings for getting the correct path. ### Description The new PUT endpoint for editing a text file in a specific language for an App repository should be created in the new TextsController which handles all interactions with text files in the new format, meaning that it's reached from the version 2 reoute URL: `designer/api/v2/{org}/{repo}/texts`. The endpoint will take a json object consisting of all the ID's and text pairs as key:value pairs as its payload. The json object should overwrite the existing text file for the language defined in the URL route. _NB: new endpoint will not sort the dictionary before storing it, if this is a desired feature it has to be done in frontend_. Endpoint is called repeatedly by some client-triggered action ( e.g. when user stops editing for some while). This includes when editing a value connected to an already existing key/ID, editing an existing key or adding a new key:value pair. All the key:value pairs will be overwritten with each PUT call to the endpoint. ### Additional Information See issue #8808 for an explanation of the conversion from old text format to new text format. ### Tasks _No response_ ### Acceptance Criterias - [ ] Implementation of EditText() endpoint - [ ] Unit/integration tests for the endpoint
non_process
put endpoint for editing text files background today text files are updated by building a new text file in the old text format with the filename format resource json the endpoint for saving the text is now defined in textcontroller as a post request to saveresource which checks the jsondata from requestbody for duplicate keys and sorts the dictionary before storing it the operations for storing overwriting the text exists in a method defined in irepository which uses a method in servicerepositorysettings for getting the correct path description the new put endpoint for editing a text file in a specific language for an app repository should be created in the new textscontroller which handles all interactions with text files in the new format meaning that it s reached from the version reoute url designer api org repo texts the endpoint will take a json object consisting of all the id s and text pairs as key value pairs as its payload the json object should overwrite the existing text file for the language defined in the url route nb new endpoint will not sort the dictionary before storing it if this is a desired feature it has to be done in frontend endpoint is called repeatedly by some client triggered action e g when user stops editing for some while this includes when editing a value connected to an already existing key id editing an existing key or adding a new key value pair all the key value pairs will be overwritten with each put call to the endpoint additional information see issue for an explanation of the conversion from old text format to new text format tasks no response acceptance criterias implementation of edittext endpoint unit integration tests for the endpoint
0
183,207
21,719,411,371
IssuesEvent
2022-05-10 21:37:41
webc-zss2/traefik
https://api.github.com/repos/webc-zss2/traefik
opened
CVE-2021-42581 (Medium) detected in ramda-0.26.1.tgz
security vulnerability
## CVE-2021-42581 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ramda-0.26.1.tgz</b></p></summary> <p>A practical functional library for JavaScript programmers.</p> <p>Library home page: <a href="https://registry.npmjs.org/ramda/-/ramda-0.26.1.tgz">https://registry.npmjs.org/ramda/-/ramda-0.26.1.tgz</a></p> <p>Path to dependency file: /webui/package.json</p> <p>Path to vulnerable library: /webui/node_modules/ramda/package.json</p> <p> Dependency Hierarchy: - eslint-plugin-mocha-6.2.1.tgz (Root Library) - :x: **ramda-0.26.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/webc-zss2/traefik/commit/5fb0a687f6f4614ef22e971903247af1c371d3c8">5fb0a687f6f4614ef22e971903247af1c371d3c8</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype poisoning in function mapObjIndexed in Ramda 0.27.0 and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object (that contains an own property "__proto__") as an argument to the function. <p>Publish Date: 2022-05-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42581>CVE-2021-42581</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42581">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42581</a></p> <p>Release Date: 2022-05-10</p> <p>Fix Resolution (ramda): 0.27.1</p> <p>Direct dependency fix Resolution (eslint-plugin-mocha): 6.3.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"eslint-plugin-mocha","packageVersion":"6.2.1","packageFilePaths":["/webui/package.json"],"isTransitiveDependency":false,"dependencyTree":"eslint-plugin-mocha:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.3.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-42581","vulnerabilityDetails":"Prototype poisoning in function mapObjIndexed in Ramda 0.27.0 and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object (that contains an own property \"__proto__\") as an argument to the function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42581","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-42581 (Medium) detected in ramda-0.26.1.tgz - ## CVE-2021-42581 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ramda-0.26.1.tgz</b></p></summary> <p>A practical functional library for JavaScript programmers.</p> <p>Library home page: <a href="https://registry.npmjs.org/ramda/-/ramda-0.26.1.tgz">https://registry.npmjs.org/ramda/-/ramda-0.26.1.tgz</a></p> <p>Path to dependency file: /webui/package.json</p> <p>Path to vulnerable library: /webui/node_modules/ramda/package.json</p> <p> Dependency Hierarchy: - eslint-plugin-mocha-6.2.1.tgz (Root Library) - :x: **ramda-0.26.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/webc-zss2/traefik/commit/5fb0a687f6f4614ef22e971903247af1c371d3c8">5fb0a687f6f4614ef22e971903247af1c371d3c8</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype poisoning in function mapObjIndexed in Ramda 0.27.0 and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object (that contains an own property "__proto__") as an argument to the function. <p>Publish Date: 2022-05-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42581>CVE-2021-42581</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42581">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-42581</a></p> <p>Release Date: 2022-05-10</p> <p>Fix Resolution (ramda): 0.27.1</p> <p>Direct dependency fix Resolution (eslint-plugin-mocha): 6.3.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"eslint-plugin-mocha","packageVersion":"6.2.1","packageFilePaths":["/webui/package.json"],"isTransitiveDependency":false,"dependencyTree":"eslint-plugin-mocha:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.3.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-42581","vulnerabilityDetails":"Prototype poisoning in function mapObjIndexed in Ramda 0.27.0 and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object (that contains an own property \"__proto__\") as an argument to the function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-42581","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in ramda tgz cve medium severity vulnerability vulnerable library ramda tgz a practical functional library for javascript programmers library home page a href path to dependency file webui package json path to vulnerable library webui node modules ramda package json dependency hierarchy eslint plugin mocha tgz root library x ramda tgz vulnerable library found in head commit a href found in base branch master vulnerability details prototype poisoning in function mapobjindexed in ramda and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object that contains an own property proto as an argument to the function publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ramda direct dependency fix resolution eslint plugin mocha check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree eslint plugin mocha isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails prototype poisoning in function mapobjindexed in ramda and earlier allows attackers to compromise integrity or availability of application via supplying a crafted object that contains an own property proto as an argument to the function vulnerabilityurl
0
337,380
10,217,029,871
IssuesEvent
2019-08-15 12:34:14
mozilla/addons-server
https://api.github.com/repos/mozilla/addons-server
closed
Radio buttons are overlapping the secondary hero module icons
component: admin tools priority: p5 state: pull request ready
Follow up for #11780 ### Describe the problem and steps to reproduce it: 1. Log in to AMO -dev with an admin account 2. Access the Secondary hero shelves module - https://addons-internal.dev.mozaws.net/en-US/admin/models/discovery/secondaryheroshelf/ 3. Open one of the available entries 4. Notice the list of icons displayed for each hero ### What happened? Most of the icons are overlapped by their corresponding radio button ### What did you expect to happen? The radio buttons should be displayed next to the icons ### Anything else we should know? - if you see any other display issues, do a hard refresh - reproduced on -dev with Fx68, Win10x64 ![image](https://user-images.githubusercontent.com/31961530/62870748-d3b67a80-bd22-11e9-85f4-032cd8f2c187.png)
1.0
Radio buttons are overlapping the secondary hero module icons - Follow up for #11780 ### Describe the problem and steps to reproduce it: 1. Log in to AMO -dev with an admin account 2. Access the Secondary hero shelves module - https://addons-internal.dev.mozaws.net/en-US/admin/models/discovery/secondaryheroshelf/ 3. Open one of the available entries 4. Notice the list of icons displayed for each hero ### What happened? Most of the icons are overlapped by their corresponding radio button ### What did you expect to happen? The radio buttons should be displayed next to the icons ### Anything else we should know? - if you see any other display issues, do a hard refresh - reproduced on -dev with Fx68, Win10x64 ![image](https://user-images.githubusercontent.com/31961530/62870748-d3b67a80-bd22-11e9-85f4-032cd8f2c187.png)
non_process
radio buttons are overlapping the secondary hero module icons follow up for describe the problem and steps to reproduce it log in to amo dev with an admin account access the secondary hero shelves module open one of the available entries notice the list of icons displayed for each hero what happened most of the icons are overlapped by their corresponding radio button what did you expect to happen the radio buttons should be displayed next to the icons anything else we should know if you see any other display issues do a hard refresh reproduced on dev with
0
7,646
10,738,544,668
IssuesEvent
2019-10-29 14:57:02
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
opened
Update What's next page to include content on updating application
Apply Process State Dept.
Who: Applicants What: View information on what happens next Why: In order to provide information about how to update and submit prior to closing Acceptance Criteria: Update the What's next page to include additional information related to updating the application, clicking submit, and all before 11:59 EST Current Screen Shot: ![image.png](https://images.zenhubusercontent.com/59ee08f1a468affe6df7cd6f/2ac98591-3ce9-4b9a-9fa6-3bb876889e58)
1.0
Update What's next page to include content on updating application - Who: Applicants What: View information on what happens next Why: In order to provide information about how to update and submit prior to closing Acceptance Criteria: Update the What's next page to include additional information related to updating the application, clicking submit, and all before 11:59 EST Current Screen Shot: ![image.png](https://images.zenhubusercontent.com/59ee08f1a468affe6df7cd6f/2ac98591-3ce9-4b9a-9fa6-3bb876889e58)
process
update what s next page to include content on updating application who applicants what view information on what happens next why in order to provide information about how to update and submit prior to closing acceptance criteria update the what s next page to include additional information related to updating the application clicking submit and all before est current screen shot
1
5,460
8,321,163,047
IssuesEvent
2018-09-25 22:44:39
GoogleCloudPlatform/google-cloud-python
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
closed
BigQuery: 'test_load_table_from_uri_autodetect' snippet flakes w/ 409 in dataset creation
api: bigquery flaky testing type: process
From: https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/8391 ```python _____________________ test_load_table_from_uri_autodetect ______________________ client = <google.cloud.bigquery.client.Client object at 0x7fa5456d77f0> to_delete = [] capsys = <_pytest.capture.CaptureFixture object at 0x7fa5456d7a58> def test_load_table_from_uri_autodetect(client, to_delete, capsys): """Load table from a GCS URI using various formats and auto-detected schema Each file format has its own tested load from URI sample. Because most of the code is common for autodetect, append, and truncate, this sample includes snippets for all supported formats but only calls a single load job. This code snippet is made up of shared code, then format-specific code, followed by more shared code. Note that only the last format in the format-specific code section will be tested in this test. """ dataset_id = 'load_table_dataset_{}'.format(_millis()) dataset = bigquery.Dataset(client.dataset(dataset_id)) > client.create_dataset(dataset) docs/snippets.py:1463: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/bigquery/client.py:305: in create_dataset method='POST', path=path, data=data) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.bigquery._http.Connection object at 0x7fa5456d7780> method = 'POST', path = '/projects/precise-truck-742/datasets' query_params = None data = '{"datasetReference": {"projectId": "precise-truck-742", "datasetId": "load_table_dataset_1537900066881"}, "labels": {}}' content_type = 'application/json', headers = None, api_base_url = None api_version = None, expect_json = True, _target_object = None def api_request(self, method, path, query_params=None, data=None, content_type=None, headers=None, api_base_url=None, api_version=None, expect_json=True, _target_object=None): """Make a request over the HTTP transport to the API. You shouldn't need to use this method, but if you plan to interact with the API using these primitives, this is the correct one to use. :type method: str :param method: The HTTP method name (ie, ``GET``, ``POST``, etc). Required. :type path: str :param path: The path to the resource (ie, ``'/b/bucket-name'``). Required. :type query_params: dict or list :param query_params: A dictionary of keys and values (or list of key-value pairs) to insert into the query string of the URL. :type data: str :param data: The data to send as the body of the request. Default is the empty string. :type content_type: str :param content_type: The proper MIME type of the data provided. Default is None. :type headers: dict :param headers: extra HTTP headers to be sent with the request. :type api_base_url: str :param api_base_url: The base URL for the API endpoint. Typically you won't have to provide this. Default is the standard API base URL. :type api_version: str :param api_version: The version of the API to call. Typically you shouldn't provide this and instead use the default for the library. Default is the latest API version supported by google-cloud-python. :type expect_json: bool :param expect_json: If True, this method will try to parse the response as JSON and raise an exception if that cannot be done. Default is True. :type _target_object: :class:`object` :param _target_object: (Optional) Protected argument to be used by library callers. This can allow custom behavior, for example, to defer an HTTP request and complete initialization of the object at a later time. :raises ~google.cloud.exceptions.GoogleCloudError: if the response code is not 200 OK. :raises ValueError: if the response content type is not JSON. :rtype: dict or str :returns: The API response payload, either as a raw string or a dictionary if the response is valid JSON. """ url = self.build_api_url(path=path, query_params=query_params, api_base_url=api_base_url, api_version=api_version) # Making the executive decision that any dictionary # data will be sent properly as JSON. if data and isinstance(data, dict): data = json.dumps(data) content_type = 'application/json' response = self._make_request( method=method, url=url, data=data, content_type=content_type, headers=headers, target_object=_target_object) if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E google.api_core.exceptions.Conflict: 409 POST https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets: Already Exists: Dataset precise-truck-742:load_table_dataset_1537900066881 ```
1.0
BigQuery: 'test_load_table_from_uri_autodetect' snippet flakes w/ 409 in dataset creation - From: https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/8391 ```python _____________________ test_load_table_from_uri_autodetect ______________________ client = <google.cloud.bigquery.client.Client object at 0x7fa5456d77f0> to_delete = [] capsys = <_pytest.capture.CaptureFixture object at 0x7fa5456d7a58> def test_load_table_from_uri_autodetect(client, to_delete, capsys): """Load table from a GCS URI using various formats and auto-detected schema Each file format has its own tested load from URI sample. Because most of the code is common for autodetect, append, and truncate, this sample includes snippets for all supported formats but only calls a single load job. This code snippet is made up of shared code, then format-specific code, followed by more shared code. Note that only the last format in the format-specific code section will be tested in this test. """ dataset_id = 'load_table_dataset_{}'.format(_millis()) dataset = bigquery.Dataset(client.dataset(dataset_id)) > client.create_dataset(dataset) docs/snippets.py:1463: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/bigquery/client.py:305: in create_dataset method='POST', path=path, data=data) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.bigquery._http.Connection object at 0x7fa5456d7780> method = 'POST', path = '/projects/precise-truck-742/datasets' query_params = None data = '{"datasetReference": {"projectId": "precise-truck-742", "datasetId": "load_table_dataset_1537900066881"}, "labels": {}}' content_type = 'application/json', headers = None, api_base_url = None api_version = None, expect_json = True, _target_object = None def api_request(self, method, path, query_params=None, data=None, content_type=None, headers=None, api_base_url=None, api_version=None, expect_json=True, _target_object=None): """Make a request over the HTTP transport to the API. You shouldn't need to use this method, but if you plan to interact with the API using these primitives, this is the correct one to use. :type method: str :param method: The HTTP method name (ie, ``GET``, ``POST``, etc). Required. :type path: str :param path: The path to the resource (ie, ``'/b/bucket-name'``). Required. :type query_params: dict or list :param query_params: A dictionary of keys and values (or list of key-value pairs) to insert into the query string of the URL. :type data: str :param data: The data to send as the body of the request. Default is the empty string. :type content_type: str :param content_type: The proper MIME type of the data provided. Default is None. :type headers: dict :param headers: extra HTTP headers to be sent with the request. :type api_base_url: str :param api_base_url: The base URL for the API endpoint. Typically you won't have to provide this. Default is the standard API base URL. :type api_version: str :param api_version: The version of the API to call. Typically you shouldn't provide this and instead use the default for the library. Default is the latest API version supported by google-cloud-python. :type expect_json: bool :param expect_json: If True, this method will try to parse the response as JSON and raise an exception if that cannot be done. Default is True. :type _target_object: :class:`object` :param _target_object: (Optional) Protected argument to be used by library callers. This can allow custom behavior, for example, to defer an HTTP request and complete initialization of the object at a later time. :raises ~google.cloud.exceptions.GoogleCloudError: if the response code is not 200 OK. :raises ValueError: if the response content type is not JSON. :rtype: dict or str :returns: The API response payload, either as a raw string or a dictionary if the response is valid JSON. """ url = self.build_api_url(path=path, query_params=query_params, api_base_url=api_base_url, api_version=api_version) # Making the executive decision that any dictionary # data will be sent properly as JSON. if data and isinstance(data, dict): data = json.dumps(data) content_type = 'application/json' response = self._make_request( method=method, url=url, data=data, content_type=content_type, headers=headers, target_object=_target_object) if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E google.api_core.exceptions.Conflict: 409 POST https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets: Already Exists: Dataset precise-truck-742:load_table_dataset_1537900066881 ```
process
bigquery test load table from uri autodetect snippet flakes w in dataset creation from python test load table from uri autodetect client to delete capsys def test load table from uri autodetect client to delete capsys load table from a gcs uri using various formats and auto detected schema each file format has its own tested load from uri sample because most of the code is common for autodetect append and truncate this sample includes snippets for all supported formats but only calls a single load job this code snippet is made up of shared code then format specific code followed by more shared code note that only the last format in the format specific code section will be tested in this test dataset id load table dataset format millis dataset bigquery dataset client dataset dataset id client create dataset dataset docs snippets py google cloud bigquery client py in create dataset method post path path data data self method post path projects precise truck datasets query params none data datasetreference projectid precise truck datasetid load table dataset labels content type application json headers none api base url none api version none expect json true target object none def api request self method path query params none data none content type none headers none api base url none api version none expect json true target object none make a request over the http transport to the api you shouldn t need to use this method but if you plan to interact with the api using these primitives this is the correct one to use type method str param method the http method name ie get post etc required type path str param path the path to the resource ie b bucket name required type query params dict or list param query params a dictionary of keys and values or list of key value pairs to insert into the query string of the url type data str param data the data to send as the body of the request default is the empty string type content type str param content type the proper mime type of the data provided default is none type headers dict param headers extra http headers to be sent with the request type api base url str param api base url the base url for the api endpoint typically you won t have to provide this default is the standard api base url type api version str param api version the version of the api to call typically you shouldn t provide this and instead use the default for the library default is the latest api version supported by google cloud python type expect json bool param expect json if true this method will try to parse the response as json and raise an exception if that cannot be done default is true type target object class object param target object optional protected argument to be used by library callers this can allow custom behavior for example to defer an http request and complete initialization of the object at a later time raises google cloud exceptions googleclouderror if the response code is not ok raises valueerror if the response content type is not json rtype dict or str returns the api response payload either as a raw string or a dictionary if the response is valid json url self build api url path path query params query params api base url api base url api version api version making the executive decision that any dictionary data will be sent properly as json if data and isinstance data dict data json dumps data content type application json response self make request method method url url data data content type content type headers headers target object target object if not response status code raise exceptions from http response response e google api core exceptions conflict post already exists dataset precise truck load table dataset
1
2,648
5,427,978,045
IssuesEvent
2017-03-03 14:50:10
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
supervisord configuration for use of zcat with multiple log files and websocket
html report log-processing websocket-server
I want to control goaccess from supervisord for real-time web reports. What should be the command? would be great and helpful if you add an example supervisord configuration. this is command with supervisord `zcat -f /var/log/apache2/access.log.* | goaccess --log-format=COMBINED -o /var/www/html/report.html --real-time-html --ws-url=server.ip -p /etc/goaccess.conf /var/log/apache2/access.log -` thanks in advance :)
1.0
supervisord configuration for use of zcat with multiple log files and websocket - I want to control goaccess from supervisord for real-time web reports. What should be the command? would be great and helpful if you add an example supervisord configuration. this is command with supervisord `zcat -f /var/log/apache2/access.log.* | goaccess --log-format=COMBINED -o /var/www/html/report.html --real-time-html --ws-url=server.ip -p /etc/goaccess.conf /var/log/apache2/access.log -` thanks in advance :)
process
supervisord configuration for use of zcat with multiple log files and websocket i want to control goaccess from supervisord for real time web reports what should be the command would be great and helpful if you add an example supervisord configuration this is command with supervisord zcat f var log access log goaccess log format combined o var www html report html real time html ws url server ip p etc goaccess conf var log access log thanks in advance
1
703,028
24,144,925,329
IssuesEvent
2022-09-21 17:51:14
freeorion/freeorion
https://api.github.com/repos/freeorion/freeorion
closed
Server crash when process new turn
category:bug component:network component:internal priority:high
Bug Report ========== <!-- Please fill in a a meaningful and concise title in the field above. It helps us to classify the issue even before reading the body and handling the issue a bit better. Examples: * Great: "Unable to select fleet in galaxy map" * Good: "Unable to click fleet" * Not really okay: "Issue with fleets" * Horrible: "HELP!!!" --> Environment ----------- <!-- You don't need to provide the information in section when they are not useful in the context of this issue. For example a crashing hosted server may not need information about the graphic card, but maybe depend on the operating system you use. When in doubt provide all information you have available. You can find the FreeOrion version number in the lower right corner of the game main screen. When stating the operating system also note the version of the operating system so, `Windows 8 Pro` instead of `Windows` or `Mac OSX Mountain Lion` instead of `OSX`. --> * **FreeOrion Version**: Custom https://github.com/o01eg/freeorion/commit/738aa62ebb8882d0ee7e17592fe44bb94ff32da1 based on 000f16b * **Operating System**: Ubuntu Linux 22.04.1 LTS Amd64 * **Fetched as** <!-- delete lines below that don't apply --> * Compiled from source Description ----------- <!-- Add a meaningful description of the bug you encountered. If a screenshot or image helps to describe the issue content feel free to create one. Also attach the log files the game creates. The log files The log files can be found on * Windows: %APPDATA%\FreeOrion * MacOSX: ~/Library/"Application Support"/FreeOrion * Linux: ${XDG_DATA_HOME:-~/.local/share}/freeorion --> Server sends notification about new turn but crashes and loaded save from previous one. Latest log lines: ``` 2022-09-18 17:52:18.582239 {0x00007f4a0d13d000} [debug] server : SerializeUniverse.cpp:261 : Universe serializing done 2022-09-18 17:52:18.582250 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - collecting data time: 2 ms 2022-09-18 17:52:18.582256 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - designs time: 3 ms 2022-09-18 17:52:18.582260 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - visibility / known destroyed or stale time: 12 ms 2022-09-18 17:52:18.582286 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - objects time: 59 ms 2022-09-18 17:52:18.582299 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:276 : Universe serializing time: 78.8 ms ``` Stacktrace: ``` Core was generated by `/usr/lib/freeorion/freeoriond --hostless --load-or-quickstart --resource.path /'. Program terminated with signal SIGABRT, Aborted. #0 0x00007f4a0f591a7c in pthread_kill () from /lib/x86_64-linux-gnu/libc.so.6 (gdb) bt #0 0x00007f4a0f591a7c in pthread_kill () from /lib/x86_64-linux-gnu/libc.so.6 #1 0x00007f4a0f53d476 in raise () from /lib/x86_64-linux-gnu/libc.so.6 #2 0x00007f4a0f5237f3 in abort () from /lib/x86_64-linux-gnu/libc.so.6 #3 0x00007f4a0f8cebfe in ?? () from /lib/x86_64-linux-gnu/libstdc++.so.6 #4 0x00007f4a0f8da28c in ?? () from /lib/x86_64-linux-gnu/libstdc++.so.6 #5 0x00007f4a0f8da2f7 in std::terminate() () from /lib/x86_64-linux-gnu/libstdc++.so.6 #6 0x00007f4a0f8da558 in __cxa_throw () from /lib/x86_64-linux-gnu/libstdc++.so.6 #7 0x000055c5e64a5480 in boost::throw_exception<boost::system::system_error>(boost::system::system_error const&) [clone .constprop.0] (e=...) at /usr/include/boost/throw_exception.hpp:165 #8 0x000055c5e657cae3 in boost::asio::detail::do_throw_error (err=..., location=<optimized out>) at /usr/include/boost/asio/detail/impl/throw_error.ipp:38 #9 0x000055c5e662a92a in boost::asio::detail::throw_error (location=<optimized out>, err=..., err=..., location=<optimized out>) at /usr/include/boost/asio/detail/throw_error.hpp:42 #10 boost::asio::basic_socket<boost::asio::ip::tcp, boost::asio::execution::any_executor<boost::asio::execution::context_as_t<boost::asio::execution_context&>, boost::asio::execution::detail::blocking::never_t<0>, boost::asio::execution::prefer_only<boost::asio::execution::detail::blocking::possibly_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::outstanding_work::tracked_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::outstanding_work::untracked_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::relationship::fork_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::relationship::continuation_t<0> > > >::remote_endpoint (this=<optimized out>) at /usr/include/boost/asio/basic_socket.hpp:1609 #11 PlayerConnection::IsLocalConnection (this=<optimized out>) at ./server/ServerNetworking.cpp:224 #12 0x000055c5e65d65c3 in ServerApp::PreCombatProcessTurns (this=0x7ffd87d2bbf0) at /usr/include/c++/11/bits/shared_ptr_base.h:1295 #13 0x000055c5e66677c2 in ProcessingTurn::react(ProcessTurn const&) [clone .constprop.0] (this=0x55c5e7307710, u=...) at ./server/ServerFSM.cpp:3634 #14 0x000055c5e6624705 in boost::statechart::custom_reaction<ProcessTurn>::react<ProcessingTurn, boost::statechart::event_base, void const*> (eventType=<optimized out>, evt=..., stt=...) at /usr/include/boost/statechart/custom_reaction.hpp:41 ... ``` Expected Result --------------- <!-- When you enter an issue please add a description of what behaviour you would expect from the game instead of the issue. --> Correct turn advance. Steps to reproduce ------------------ <!-- If the error only manifests itself after doing a certain number of actions please add a list of steps to reproduce the bug. If the error only occurs in a certain game state please be sure to add a save game to the issue. To upload save games please pack them into a Zip archive (*.zip) first. --> Not sure if it reproducible, I suppose player disconnected during turn processing.
1.0
Server crash when process new turn - Bug Report ========== <!-- Please fill in a a meaningful and concise title in the field above. It helps us to classify the issue even before reading the body and handling the issue a bit better. Examples: * Great: "Unable to select fleet in galaxy map" * Good: "Unable to click fleet" * Not really okay: "Issue with fleets" * Horrible: "HELP!!!" --> Environment ----------- <!-- You don't need to provide the information in section when they are not useful in the context of this issue. For example a crashing hosted server may not need information about the graphic card, but maybe depend on the operating system you use. When in doubt provide all information you have available. You can find the FreeOrion version number in the lower right corner of the game main screen. When stating the operating system also note the version of the operating system so, `Windows 8 Pro` instead of `Windows` or `Mac OSX Mountain Lion` instead of `OSX`. --> * **FreeOrion Version**: Custom https://github.com/o01eg/freeorion/commit/738aa62ebb8882d0ee7e17592fe44bb94ff32da1 based on 000f16b * **Operating System**: Ubuntu Linux 22.04.1 LTS Amd64 * **Fetched as** <!-- delete lines below that don't apply --> * Compiled from source Description ----------- <!-- Add a meaningful description of the bug you encountered. If a screenshot or image helps to describe the issue content feel free to create one. Also attach the log files the game creates. The log files The log files can be found on * Windows: %APPDATA%\FreeOrion * MacOSX: ~/Library/"Application Support"/FreeOrion * Linux: ${XDG_DATA_HOME:-~/.local/share}/freeorion --> Server sends notification about new turn but crashes and loaded save from previous one. Latest log lines: ``` 2022-09-18 17:52:18.582239 {0x00007f4a0d13d000} [debug] server : SerializeUniverse.cpp:261 : Universe serializing done 2022-09-18 17:52:18.582250 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - collecting data time: 2 ms 2022-09-18 17:52:18.582256 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - designs time: 3 ms 2022-09-18 17:52:18.582260 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - visibility / known destroyed or stale time: 12 ms 2022-09-18 17:52:18.582286 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:266 : Universe serializing - objects time: 59 ms 2022-09-18 17:52:18.582299 {0x00007f4a0d13d000} [debug] timer : ScopedTimer.cpp:276 : Universe serializing time: 78.8 ms ``` Stacktrace: ``` Core was generated by `/usr/lib/freeorion/freeoriond --hostless --load-or-quickstart --resource.path /'. Program terminated with signal SIGABRT, Aborted. #0 0x00007f4a0f591a7c in pthread_kill () from /lib/x86_64-linux-gnu/libc.so.6 (gdb) bt #0 0x00007f4a0f591a7c in pthread_kill () from /lib/x86_64-linux-gnu/libc.so.6 #1 0x00007f4a0f53d476 in raise () from /lib/x86_64-linux-gnu/libc.so.6 #2 0x00007f4a0f5237f3 in abort () from /lib/x86_64-linux-gnu/libc.so.6 #3 0x00007f4a0f8cebfe in ?? () from /lib/x86_64-linux-gnu/libstdc++.so.6 #4 0x00007f4a0f8da28c in ?? () from /lib/x86_64-linux-gnu/libstdc++.so.6 #5 0x00007f4a0f8da2f7 in std::terminate() () from /lib/x86_64-linux-gnu/libstdc++.so.6 #6 0x00007f4a0f8da558 in __cxa_throw () from /lib/x86_64-linux-gnu/libstdc++.so.6 #7 0x000055c5e64a5480 in boost::throw_exception<boost::system::system_error>(boost::system::system_error const&) [clone .constprop.0] (e=...) at /usr/include/boost/throw_exception.hpp:165 #8 0x000055c5e657cae3 in boost::asio::detail::do_throw_error (err=..., location=<optimized out>) at /usr/include/boost/asio/detail/impl/throw_error.ipp:38 #9 0x000055c5e662a92a in boost::asio::detail::throw_error (location=<optimized out>, err=..., err=..., location=<optimized out>) at /usr/include/boost/asio/detail/throw_error.hpp:42 #10 boost::asio::basic_socket<boost::asio::ip::tcp, boost::asio::execution::any_executor<boost::asio::execution::context_as_t<boost::asio::execution_context&>, boost::asio::execution::detail::blocking::never_t<0>, boost::asio::execution::prefer_only<boost::asio::execution::detail::blocking::possibly_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::outstanding_work::tracked_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::outstanding_work::untracked_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::relationship::fork_t<0> >, boost::asio::execution::prefer_only<boost::asio::execution::detail::relationship::continuation_t<0> > > >::remote_endpoint (this=<optimized out>) at /usr/include/boost/asio/basic_socket.hpp:1609 #11 PlayerConnection::IsLocalConnection (this=<optimized out>) at ./server/ServerNetworking.cpp:224 #12 0x000055c5e65d65c3 in ServerApp::PreCombatProcessTurns (this=0x7ffd87d2bbf0) at /usr/include/c++/11/bits/shared_ptr_base.h:1295 #13 0x000055c5e66677c2 in ProcessingTurn::react(ProcessTurn const&) [clone .constprop.0] (this=0x55c5e7307710, u=...) at ./server/ServerFSM.cpp:3634 #14 0x000055c5e6624705 in boost::statechart::custom_reaction<ProcessTurn>::react<ProcessingTurn, boost::statechart::event_base, void const*> (eventType=<optimized out>, evt=..., stt=...) at /usr/include/boost/statechart/custom_reaction.hpp:41 ... ``` Expected Result --------------- <!-- When you enter an issue please add a description of what behaviour you would expect from the game instead of the issue. --> Correct turn advance. Steps to reproduce ------------------ <!-- If the error only manifests itself after doing a certain number of actions please add a list of steps to reproduce the bug. If the error only occurs in a certain game state please be sure to add a save game to the issue. To upload save games please pack them into a Zip archive (*.zip) first. --> Not sure if it reproducible, I suppose player disconnected during turn processing.
non_process
server crash when process new turn bug report please fill in a a meaningful and concise title in the field above it helps us to classify the issue even before reading the body and handling the issue a bit better examples great unable to select fleet in galaxy map good unable to click fleet not really okay issue with fleets horrible help environment you don t need to provide the information in section when they are not useful in the context of this issue for example a crashing hosted server may not need information about the graphic card but maybe depend on the operating system you use when in doubt provide all information you have available you can find the freeorion version number in the lower right corner of the game main screen when stating the operating system also note the version of the operating system so windows pro instead of windows or mac osx mountain lion instead of osx freeorion version custom based on operating system ubuntu linux lts fetched as compiled from source description add a meaningful description of the bug you encountered if a screenshot or image helps to describe the issue content feel free to create one also attach the log files the game creates the log files the log files can be found on windows appdata freeorion macosx library application support freeorion linux xdg data home local share freeorion server sends notification about new turn but crashes and loaded save from previous one latest log lines server serializeuniverse cpp universe serializing done timer scopedtimer cpp universe serializing collecting data time ms timer scopedtimer cpp universe serializing designs time ms timer scopedtimer cpp universe serializing visibility known destroyed or stale time ms timer scopedtimer cpp universe serializing objects time ms timer scopedtimer cpp universe serializing time ms stacktrace core was generated by usr lib freeorion freeoriond hostless load or quickstart resource path program terminated with signal sigabrt aborted in pthread kill from lib linux gnu libc so gdb bt in pthread kill from lib linux gnu libc so in raise from lib linux gnu libc so in abort from lib linux gnu libc so in from lib linux gnu libstdc so in from lib linux gnu libstdc so in std terminate from lib linux gnu libstdc so in cxa throw from lib linux gnu libstdc so in boost throw exception boost system system error const e at usr include boost throw exception hpp in boost asio detail do throw error err location at usr include boost asio detail impl throw error ipp in boost asio detail throw error location err err location at usr include boost asio detail throw error hpp boost asio basic socket boost asio execution detail blocking never t boost asio execution prefer only boost asio execution prefer only boost asio execution prefer only boost asio execution prefer only boost asio execution prefer only remote endpoint this at usr include boost asio basic socket hpp playerconnection islocalconnection this at server servernetworking cpp in serverapp precombatprocessturns this at usr include c bits shared ptr base h in processingturn react processturn const this u at server serverfsm cpp in boost statechart custom reaction react eventtype evt stt at usr include boost statechart custom reaction hpp expected result when you enter an issue please add a description of what behaviour you would expect from the game instead of the issue correct turn advance steps to reproduce if the error only manifests itself after doing a certain number of actions please add a list of steps to reproduce the bug if the error only occurs in a certain game state please be sure to add a save game to the issue to upload save games please pack them into a zip archive zip first not sure if it reproducible i suppose player disconnected during turn processing
0
15,680
19,847,727,221
IssuesEvent
2022-01-21 08:48:29
ooi-data/RS01SBPD-DP01A-03-FLCDRA102-recovered_wfp-dpc_flcdrtd_instrument_recovered
https://api.github.com/repos/ooi-data/RS01SBPD-DP01A-03-FLCDRA102-recovered_wfp-dpc_flcdrtd_instrument_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T08:48:28.842309. ## Details Flow name: `RS01SBPD-DP01A-03-FLCDRA102-recovered_wfp-dpc_flcdrtd_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T08:48:28.842309. ## Details Flow name: `RS01SBPD-DP01A-03-FLCDRA102-recovered_wfp-dpc_flcdrtd_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered wfp dpc flcdrtd instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
20,979
4,651,530,675
IssuesEvent
2016-10-03 10:29:31
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Documentation about migrate WAR to executable Spring Boot is misleading
documentation
Documentation [howto-convert-an-existing-application-to-spring-boot](http://docs.spring.io/spring-boot/docs/1.3.8.RELEASE/reference/htmlsingle/#howto-convert-an-existing-application-to-spring-boot ) suggests that if you want to migrate your WAR to executable Spring Boot application in first step `SpringBootServletInitializer ` should be implemented with method `configure`. In second step `main` method should be added. It looks that recommended way is implement method `configure`, but [the method is not called in standalone application] (http://stackoverflow.com/questions/39729012/springbootservletinitializerconfigure-is-not-called-for-executable-war) (when application starts from `main`, not as WAR on servlet container)
1.0
Documentation about migrate WAR to executable Spring Boot is misleading - Documentation [howto-convert-an-existing-application-to-spring-boot](http://docs.spring.io/spring-boot/docs/1.3.8.RELEASE/reference/htmlsingle/#howto-convert-an-existing-application-to-spring-boot ) suggests that if you want to migrate your WAR to executable Spring Boot application in first step `SpringBootServletInitializer ` should be implemented with method `configure`. In second step `main` method should be added. It looks that recommended way is implement method `configure`, but [the method is not called in standalone application] (http://stackoverflow.com/questions/39729012/springbootservletinitializerconfigure-is-not-called-for-executable-war) (when application starts from `main`, not as WAR on servlet container)
non_process
documentation about migrate war to executable spring boot is misleading documentation suggests that if you want to migrate your war to executable spring boot application in first step springbootservletinitializer should be implemented with method configure in second step main method should be added it looks that recommended way is implement method configure but when application starts from main not as war on servlet container
0
7,411
10,533,512,238
IssuesEvent
2019-10-01 13:12:19
ericadamski/alphabet-keys
https://api.github.com/repos/ericadamski/alphabet-keys
closed
Add a counting game
enhancement hacktoberfest help wanted 👩‍💻in process
It would be awesome to have more games that can teach things like counting! Something like, a number of the same emoji pops up and the user has to enter the number they see! Really terrible text example: 🍎 🍎 🍎 3 - This will need a new menu screen to pick different games and a new input and validation flow.
1.0
Add a counting game - It would be awesome to have more games that can teach things like counting! Something like, a number of the same emoji pops up and the user has to enter the number they see! Really terrible text example: 🍎 🍎 🍎 3 - This will need a new menu screen to pick different games and a new input and validation flow.
process
add a counting game it would be awesome to have more games that can teach things like counting something like a number of the same emoji pops up and the user has to enter the number they see really terrible text example 🍎 🍎 🍎 this will need a new menu screen to pick different games and a new input and validation flow
1
53,780
6,344,280,265
IssuesEvent
2017-07-27 19:32:21
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
closed
Manual test run on Windows x64 for 0.18.x (Beta Channel)
OS/Windows release-notes/exclude tests
## Per release specialty tests - [x] Chrome-extension link shown instead of about:safebrowsing. ([#10143](https://github.com/brave/browser-laptop/issues/10143)) - [x] ModalDialog lacks margin-top: 50px. ([#10138](https://github.com/brave/browser-laptop/issues/10138)) - [x] LastPass won't stop asking for 2FA code. ([#9259](https://github.com/brave/browser-laptop/issues/9259)) - [x] Pocket and LastPass home page aren't loaded after logging into account. ([#9117](https://github.com/brave/browser-laptop/issues/9117)) - [x] Log in pop-up window / tab should be closed on Pocket. ([#9113](https://github.com/brave/browser-laptop/issues/9113)) - [x] Brave Fails to "Remember this decision". ([#9076](https://github.com/brave/browser-laptop/issues/9076)) ## Installer 1. [x] Check that installer is close to the size of last release. 2. [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave.app/` and make sure it returns `accepted`. If Windows right click on the installer exe and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window. 3. [x] Check Brave, muon, and libchromiumcontent version in About and make sure it is EXACTLY as expected. ## Last changeset test 1. [x] Test what is covered by the last changeset (you can find this by clicking on the SHA in about:brave). ## Widevine/Netflix test 1. [x] Test that you can log into Netflix and start a show. ## Ledger 1. [x] Create a wallet with a value other than $5 selected in the monthly budget dropdown. Click on the 'Add Funds' button and check that Coinbase transactions are blocked. 2. [x] Remove all `ledger-*.json` files from `~/Library/Application\ Support/Brave/`. Go to the Payments tab in about:preferences, enable payments, click on `create wallet`. Check that the `add funds` button appears after a wallet is created. 3. [ ] Click on `add funds` and verify that adding funds through Coinbase increases the account balance. 4. [ ] Repeat the step above but add funds by scanning the QR code in a mobile bitcoin app instead of through Coinbase. 5. [x] Visit nytimes.com for a few seconds and make sure it shows up in the Payments table. 6. [x] Go to https://jsfiddle.net/LnwtLckc/5/ and click the register button. In the Payments tab, click `add funds`. Verify that the `transfer funds` button is visible and that clicking on `transfer funds` opens a jsfiddle URL in a new tab. 7. [x] Go to https://jsfiddle.net/LnwtLckc/5/ and click `unregister`. Verify that the `transfer funds` button no longer appears in the `add funds` modal. 8. [x] Check that disabling payments and enabling them again does not lose state. ## Sync 1. [x] Verify you are able to sync two devices using the secret code 2. [x] Visit a site on device 1 and change shield setting, ensure that the saved site preference is synced to device 2 3. [x] Enable Browsing history sync on device 1, ensure the history is shown on device 2 4. [x] Import/Add bookmarks on device 1, ensure it is synced on device 2 5. [x] Ensure imported bookmark folder structure is maintained on device 2 6. [x] Ensure bookmark favicons are shown after sync ## Data 1. [x] Make sure that data from the last version appears in the new version OK. 2. [x] Test that the previous version's cookies are preserved in the next version. ## About pages 1. [x] Test that about:adblock loads 2. [x] Test that about:autofill loads 3. [x] Test that about:bookmarks loads bookmarks 4. [x] Test that about:downloads loads downloads 5. [x] Test that about:extensions loads 6. [x] Test that about:history loads history 7. [x] Test that about:passwords loads 8. [x] Test that about:styles loads 9. [x] Test that about:welcome loads 10. [x] Test that about:preferences changing a preference takes effect right away 11. [x] Test that about:preferences language change takes effect on re-start ## Bookmarks 1. [x] Test that creating a bookmark on the bookmarks toolbar with the star button works 2. [x] Test that creating a bookmark on the bookmarks toolbar by dragging the un/lock icon works 3. [x] Test that creating a bookmark folder on the bookmarks toolbar works 4. [x] Test that moving a bookmark into a folder by drag and drop on the bookmarks folder works 5. [x] Test that clicking a bookmark in the toolbar loads the bookmark. 6. [x] Test that clicking a bookmark in a bookmark toolbar folder loads the bookmark. ## Context menus 1. [x] Make sure context menu items in the URL bar work 2. [x] Make sure context menu items on content work with no selected text. 3. [x] Make sure context menu items on content work with selected text. 4. [x] Make sure context menu items on content work inside an editable control on `about:styles` (input, textarea, or contenteditable). ## Find on page 1. [x] Ensure search box is shown with shortcut 2. [x] Test successful find 3. [x] Test forward and backward find navigation 4. [x] Test failed find shows 0 results 5. [x] Test match case find ## Geolocation 1. [x] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation works ## Site hacks 1. [x] Test https://www.twitch.tv/adobe sub-page loads a video and you can play it ## Downloads 1. [x] Test downloading a file works and that all actions on the download item works. ## Fullscreen 1. [x] Test that entering full screen window works View -> Toggle Full Screen. And exit back (Not Esc). 2. [x] Test that entering HTML5 full screen works. And Esc to go back. (youtube.com) ## Tabs, Pinning and Tear off tabs 1. [x] Test that tabs are pinnable 2. [x] Test that tabs are unpinnable 3. [x] Test that tabs are draggable to same tabset 4. [x] Test that tabs are draggable to alternate tabset 5. [x] Test that tabs can be teared off into a new window 6. [x] Test that you are able to reattach a tab that is teared off into a new window 7. [x] Test that tab pages can be closed 8. [x] Test that tab pages can be muted ## Zoom 1. [x] Test zoom in / out shortcut works 2. [x] Test hamburger menu zooms. 3. [x] Test zoom saved when you close the browser and restore on a single site. 4. [x] Test zoom saved when you navigate within a single origin site. 5. [x] Test that navigating to a different origin resets the zoom ## Bravery settings 1. [x] Check that HTTPS Everywhere works by loading https://https-everywhere.badssl.com/ 2. [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/ 3. [x] Check that ad replacement works on http://slashdot.org 4. [x] Check that toggling to blocking and allow ads works as expected. 5. [x] Test that clicking through a cert error in https://badssl.com/ works. 6. [x] Test that Safe Browsing works (http://downloadme.org/) 7. [x] Turning Safe Browsing off and shields off both disable safe browsing for http://downloadme.org/. 8. [x] Visit https://brianbondy.com/ and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work. 9. [x] Test that about:preferences default Bravery settings take effect on pages with no site settings. 10. [x] Test that turning on fingerprinting protection in about:preferences shows 3 fingerprints blocked at https://jsfiddle.net/bkf50r8v/13/. Test that turning it off in the Bravery menu shows 0 fingerprints blocked. 11. [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked. 12. [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ when fingerprinting protection is on. 13. [x] Test that browser is not detected on https://extensions.inrialpes.fr/brave/ ## Content tests 1. [x] Go to https://brianbondy.com/ and click on the twitter icon on the top right. Test that context menus work in the new twitter tab. 2. [x] Load twitter and click on a tweet so the popup div shows. Click to dismiss and repeat with another div. Make sure it shows. 3. [x] Go to http://www.bennish.net/web-notifications.html and test that clicking on 'Show' pops up a notification asking for permission. Make sure that clicking 'Deny' leads to no notifications being shown. 4. [x] Go to https://trac.torproject.org/projects/tor/login and make sure that the password can be saved. Make sure the saved password shows up in `about:passwords`. Then reload https://trac.torproject.org/projects/tor/login and make sure the password is autofilled. 5. [x] Open `about:styles` and type some misspellings on a textbox, make sure they are underlined. 6. [x] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text. 7. [x] Make sure that Command + Click (Control + Click on Windows, Control + Click on Ubuntu) on a link opens a new tab but does NOT switch to it. Click on it and make sure it is already loaded. 8. [x] Open an email on http://mail.google.com/ or inbox.google.com and click on a link. Make sure it works. 9. [x] Test that PDF is loaded at http://www.orimi.com/pdf-test.pdf 10. [x] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run). ## Flash tests 1. [x] Turn on Flash in about:preferences#security. Test that clicking on 'Install Flash' banner on myspace.com shows a notification to allow Flash and that the banner disappears when 'Allow' is clicked. 2. [x] Test that flash placeholder appears on http://www.homestarrunner.com ## Autofill tests 1. [x] Test that autofill works on http://www.roboform.com/filling-test-all-fields ## Session storage Do not forget to make a backup of your entire `~/Library/Application\ Support/Brave` folder. 1. [x] Temporarily move away your `~/Library/Application\ Support/Brave/session-store-1` and test that clean session storage works. (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) 2. [x] Test that windows and tabs restore when closed, including active tab. 3. [x] Move away your entire `~/Library/Application\ Support/Brave` folder (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) ## Cookie and Cache 1. [x] Make a backup of your profile, turn on all clearing in preferences and shut down. Make sure when you bring the browser back up everything is gone that is specified. 2. [x] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the Evercookie site does not remember the old evercookie value. ## Update tests 1. [ ] Test that updating using `BRAVE_UPDATE_VERSION=0.8.3` env variable works correctly.
1.0
Manual test run on Windows x64 for 0.18.x (Beta Channel) - ## Per release specialty tests - [x] Chrome-extension link shown instead of about:safebrowsing. ([#10143](https://github.com/brave/browser-laptop/issues/10143)) - [x] ModalDialog lacks margin-top: 50px. ([#10138](https://github.com/brave/browser-laptop/issues/10138)) - [x] LastPass won't stop asking for 2FA code. ([#9259](https://github.com/brave/browser-laptop/issues/9259)) - [x] Pocket and LastPass home page aren't loaded after logging into account. ([#9117](https://github.com/brave/browser-laptop/issues/9117)) - [x] Log in pop-up window / tab should be closed on Pocket. ([#9113](https://github.com/brave/browser-laptop/issues/9113)) - [x] Brave Fails to "Remember this decision". ([#9076](https://github.com/brave/browser-laptop/issues/9076)) ## Installer 1. [x] Check that installer is close to the size of last release. 2. [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave.app/` and make sure it returns `accepted`. If Windows right click on the installer exe and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window. 3. [x] Check Brave, muon, and libchromiumcontent version in About and make sure it is EXACTLY as expected. ## Last changeset test 1. [x] Test what is covered by the last changeset (you can find this by clicking on the SHA in about:brave). ## Widevine/Netflix test 1. [x] Test that you can log into Netflix and start a show. ## Ledger 1. [x] Create a wallet with a value other than $5 selected in the monthly budget dropdown. Click on the 'Add Funds' button and check that Coinbase transactions are blocked. 2. [x] Remove all `ledger-*.json` files from `~/Library/Application\ Support/Brave/`. Go to the Payments tab in about:preferences, enable payments, click on `create wallet`. Check that the `add funds` button appears after a wallet is created. 3. [ ] Click on `add funds` and verify that adding funds through Coinbase increases the account balance. 4. [ ] Repeat the step above but add funds by scanning the QR code in a mobile bitcoin app instead of through Coinbase. 5. [x] Visit nytimes.com for a few seconds and make sure it shows up in the Payments table. 6. [x] Go to https://jsfiddle.net/LnwtLckc/5/ and click the register button. In the Payments tab, click `add funds`. Verify that the `transfer funds` button is visible and that clicking on `transfer funds` opens a jsfiddle URL in a new tab. 7. [x] Go to https://jsfiddle.net/LnwtLckc/5/ and click `unregister`. Verify that the `transfer funds` button no longer appears in the `add funds` modal. 8. [x] Check that disabling payments and enabling them again does not lose state. ## Sync 1. [x] Verify you are able to sync two devices using the secret code 2. [x] Visit a site on device 1 and change shield setting, ensure that the saved site preference is synced to device 2 3. [x] Enable Browsing history sync on device 1, ensure the history is shown on device 2 4. [x] Import/Add bookmarks on device 1, ensure it is synced on device 2 5. [x] Ensure imported bookmark folder structure is maintained on device 2 6. [x] Ensure bookmark favicons are shown after sync ## Data 1. [x] Make sure that data from the last version appears in the new version OK. 2. [x] Test that the previous version's cookies are preserved in the next version. ## About pages 1. [x] Test that about:adblock loads 2. [x] Test that about:autofill loads 3. [x] Test that about:bookmarks loads bookmarks 4. [x] Test that about:downloads loads downloads 5. [x] Test that about:extensions loads 6. [x] Test that about:history loads history 7. [x] Test that about:passwords loads 8. [x] Test that about:styles loads 9. [x] Test that about:welcome loads 10. [x] Test that about:preferences changing a preference takes effect right away 11. [x] Test that about:preferences language change takes effect on re-start ## Bookmarks 1. [x] Test that creating a bookmark on the bookmarks toolbar with the star button works 2. [x] Test that creating a bookmark on the bookmarks toolbar by dragging the un/lock icon works 3. [x] Test that creating a bookmark folder on the bookmarks toolbar works 4. [x] Test that moving a bookmark into a folder by drag and drop on the bookmarks folder works 5. [x] Test that clicking a bookmark in the toolbar loads the bookmark. 6. [x] Test that clicking a bookmark in a bookmark toolbar folder loads the bookmark. ## Context menus 1. [x] Make sure context menu items in the URL bar work 2. [x] Make sure context menu items on content work with no selected text. 3. [x] Make sure context menu items on content work with selected text. 4. [x] Make sure context menu items on content work inside an editable control on `about:styles` (input, textarea, or contenteditable). ## Find on page 1. [x] Ensure search box is shown with shortcut 2. [x] Test successful find 3. [x] Test forward and backward find navigation 4. [x] Test failed find shows 0 results 5. [x] Test match case find ## Geolocation 1. [x] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation works ## Site hacks 1. [x] Test https://www.twitch.tv/adobe sub-page loads a video and you can play it ## Downloads 1. [x] Test downloading a file works and that all actions on the download item works. ## Fullscreen 1. [x] Test that entering full screen window works View -> Toggle Full Screen. And exit back (Not Esc). 2. [x] Test that entering HTML5 full screen works. And Esc to go back. (youtube.com) ## Tabs, Pinning and Tear off tabs 1. [x] Test that tabs are pinnable 2. [x] Test that tabs are unpinnable 3. [x] Test that tabs are draggable to same tabset 4. [x] Test that tabs are draggable to alternate tabset 5. [x] Test that tabs can be teared off into a new window 6. [x] Test that you are able to reattach a tab that is teared off into a new window 7. [x] Test that tab pages can be closed 8. [x] Test that tab pages can be muted ## Zoom 1. [x] Test zoom in / out shortcut works 2. [x] Test hamburger menu zooms. 3. [x] Test zoom saved when you close the browser and restore on a single site. 4. [x] Test zoom saved when you navigate within a single origin site. 5. [x] Test that navigating to a different origin resets the zoom ## Bravery settings 1. [x] Check that HTTPS Everywhere works by loading https://https-everywhere.badssl.com/ 2. [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/ 3. [x] Check that ad replacement works on http://slashdot.org 4. [x] Check that toggling to blocking and allow ads works as expected. 5. [x] Test that clicking through a cert error in https://badssl.com/ works. 6. [x] Test that Safe Browsing works (http://downloadme.org/) 7. [x] Turning Safe Browsing off and shields off both disable safe browsing for http://downloadme.org/. 8. [x] Visit https://brianbondy.com/ and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work. 9. [x] Test that about:preferences default Bravery settings take effect on pages with no site settings. 10. [x] Test that turning on fingerprinting protection in about:preferences shows 3 fingerprints blocked at https://jsfiddle.net/bkf50r8v/13/. Test that turning it off in the Bravery menu shows 0 fingerprints blocked. 11. [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked. 12. [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ when fingerprinting protection is on. 13. [x] Test that browser is not detected on https://extensions.inrialpes.fr/brave/ ## Content tests 1. [x] Go to https://brianbondy.com/ and click on the twitter icon on the top right. Test that context menus work in the new twitter tab. 2. [x] Load twitter and click on a tweet so the popup div shows. Click to dismiss and repeat with another div. Make sure it shows. 3. [x] Go to http://www.bennish.net/web-notifications.html and test that clicking on 'Show' pops up a notification asking for permission. Make sure that clicking 'Deny' leads to no notifications being shown. 4. [x] Go to https://trac.torproject.org/projects/tor/login and make sure that the password can be saved. Make sure the saved password shows up in `about:passwords`. Then reload https://trac.torproject.org/projects/tor/login and make sure the password is autofilled. 5. [x] Open `about:styles` and type some misspellings on a textbox, make sure they are underlined. 6. [x] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text. 7. [x] Make sure that Command + Click (Control + Click on Windows, Control + Click on Ubuntu) on a link opens a new tab but does NOT switch to it. Click on it and make sure it is already loaded. 8. [x] Open an email on http://mail.google.com/ or inbox.google.com and click on a link. Make sure it works. 9. [x] Test that PDF is loaded at http://www.orimi.com/pdf-test.pdf 10. [x] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run). ## Flash tests 1. [x] Turn on Flash in about:preferences#security. Test that clicking on 'Install Flash' banner on myspace.com shows a notification to allow Flash and that the banner disappears when 'Allow' is clicked. 2. [x] Test that flash placeholder appears on http://www.homestarrunner.com ## Autofill tests 1. [x] Test that autofill works on http://www.roboform.com/filling-test-all-fields ## Session storage Do not forget to make a backup of your entire `~/Library/Application\ Support/Brave` folder. 1. [x] Temporarily move away your `~/Library/Application\ Support/Brave/session-store-1` and test that clean session storage works. (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) 2. [x] Test that windows and tabs restore when closed, including active tab. 3. [x] Move away your entire `~/Library/Application\ Support/Brave` folder (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu) ## Cookie and Cache 1. [x] Make a backup of your profile, turn on all clearing in preferences and shut down. Make sure when you bring the browser back up everything is gone that is specified. 2. [x] Go to http://samy.pl/evercookie/ and set an evercookie. Check that going to prefs, clearing site data and cache, and going back to the Evercookie site does not remember the old evercookie value. ## Update tests 1. [ ] Test that updating using `BRAVE_UPDATE_VERSION=0.8.3` env variable works correctly.
non_process
manual test run on windows for x beta channel per release specialty tests chrome extension link shown instead of about safebrowsing modaldialog lacks margin top lastpass won t stop asking for code pocket and lastpass home page aren t loaded after logging into account log in pop up window tab should be closed on pocket brave fails to remember this decision installer check that installer is close to the size of last release check signature if os run spctl assess verbose applications brave app and make sure it returns accepted if windows right click on the installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window check brave muon and libchromiumcontent version in about and make sure it is exactly as expected last changeset test test what is covered by the last changeset you can find this by clicking on the sha in about brave widevine netflix test test that you can log into netflix and start a show ledger create a wallet with a value other than selected in the monthly budget dropdown click on the add funds button and check that coinbase transactions are blocked remove all ledger json files from library application support brave go to the payments tab in about preferences enable payments click on create wallet check that the add funds button appears after a wallet is created click on add funds and verify that adding funds through coinbase increases the account balance repeat the step above but add funds by scanning the qr code in a mobile bitcoin app instead of through coinbase visit nytimes com for a few seconds and make sure it shows up in the payments table go to and click the register button in the payments tab click add funds verify that the transfer funds button is visible and that clicking on transfer funds opens a jsfiddle url in a new tab go to and click unregister verify that the transfer funds button no longer appears in the add funds modal check that disabling payments and enabling them again does not lose state sync verify you are able to sync two devices using the secret code visit a site on device and change shield setting ensure that the saved site preference is synced to device enable browsing history sync on device ensure the history is shown on device import add bookmarks on device ensure it is synced on device ensure imported bookmark folder structure is maintained on device ensure bookmark favicons are shown after sync data make sure that data from the last version appears in the new version ok test that the previous version s cookies are preserved in the next version about pages test that about adblock loads test that about autofill loads test that about bookmarks loads bookmarks test that about downloads loads downloads test that about extensions loads test that about history loads history test that about passwords loads test that about styles loads test that about welcome loads test that about preferences changing a preference takes effect right away test that about preferences language change takes effect on re start bookmarks test that creating a bookmark on the bookmarks toolbar with the star button works test that creating a bookmark on the bookmarks toolbar by dragging the un lock icon works test that creating a bookmark folder on the bookmarks toolbar works test that moving a bookmark into a folder by drag and drop on the bookmarks folder works test that clicking a bookmark in the toolbar loads the bookmark test that clicking a bookmark in a bookmark toolbar folder loads the bookmark context menus make sure context menu items in the url bar work make sure context menu items on content work with no selected text make sure context menu items on content work with selected text make sure context menu items on content work inside an editable control on about styles input textarea or contenteditable find on page ensure search box is shown with shortcut test successful find test forward and backward find navigation test failed find shows results test match case find geolocation check that works site hacks test sub page loads a video and you can play it downloads test downloading a file works and that all actions on the download item works fullscreen test that entering full screen window works view toggle full screen and exit back not esc test that entering full screen works and esc to go back youtube com tabs pinning and tear off tabs test that tabs are pinnable test that tabs are unpinnable test that tabs are draggable to same tabset test that tabs are draggable to alternate tabset test that tabs can be teared off into a new window test that you are able to reattach a tab that is teared off into a new window test that tab pages can be closed test that tab pages can be muted zoom test zoom in out shortcut works test hamburger menu zooms test zoom saved when you close the browser and restore on a single site test zoom saved when you navigate within a single origin site test that navigating to a different origin resets the zoom bravery settings check that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to check that ad replacement works on check that toggling to blocking and allow ads works as expected test that clicking through a cert error in works test that safe browsing works turning safe browsing off and shields off both disable safe browsing for visit and then turn on script blocking nothing should load allow it from the script blocking ui in the url bar and it should work test that about preferences default bravery settings take effect on pages with no site settings test that turning on fingerprinting protection in about preferences shows fingerprints blocked at test that turning it off in the bravery menu shows fingerprints blocked test that party storage results are blank at when party cookies are blocked and not blank when party cookies are unblocked test that audio fingerprint is blocked at when fingerprinting protection is on test that browser is not detected on content tests go to and click on the twitter icon on the top right test that context menus work in the new twitter tab load twitter and click on a tweet so the popup div shows click to dismiss and repeat with another div make sure it shows go to and test that clicking on show pops up a notification asking for permission make sure that clicking deny leads to no notifications being shown go to and make sure that the password can be saved make sure the saved password shows up in about passwords then reload and make sure the password is autofilled open about styles and type some misspellings on a textbox make sure they are underlined make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text make sure that command click control click on windows control click on ubuntu on a link opens a new tab but does not switch to it click on it and make sure it is already loaded open an email on or inbox google com and click on a link make sure it works test that pdf is loaded at test that shows up as grey not red no mixed content scripts are run flash tests turn on flash in about preferences security test that clicking on install flash banner on myspace com shows a notification to allow flash and that the banner disappears when allow is clicked test that flash placeholder appears on autofill tests test that autofill works on session storage do not forget to make a backup of your entire library application support brave folder temporarily move away your library application support brave session store and test that clean session storage works appdata brave in windows config brave in ubuntu test that windows and tabs restore when closed including active tab move away your entire library application support brave folder appdata brave in windows config brave in ubuntu cookie and cache make a backup of your profile turn on all clearing in preferences and shut down make sure when you bring the browser back up everything is gone that is specified go to and set an evercookie check that going to prefs clearing site data and cache and going back to the evercookie site does not remember the old evercookie value update tests test that updating using brave update version env variable works correctly
0
344,737
30,756,530,359
IssuesEvent
2023-07-29 05:38:58
TheUpperPart/leaguehub-backend
https://api.github.com/repos/TheUpperPart/leaguehub-backend
closed
Test: Feat#79 테스트
test
## 💎 개발할 기능 <!-- 어떤 기능을 구현할지 알려주세요. --> - [x] 대회에 참가한 플레이어에 대한 reject 테스트 - [x] 관전자 플레이어 조회 테스트 - [x] 관리자 부여 테스트 ## 📖 참고 사항 <!-- 레퍼런스, 스크린샷 등을 넣어 주세요. -->
1.0
Test: Feat#79 테스트 - ## 💎 개발할 기능 <!-- 어떤 기능을 구현할지 알려주세요. --> - [x] 대회에 참가한 플레이어에 대한 reject 테스트 - [x] 관전자 플레이어 조회 테스트 - [x] 관리자 부여 테스트 ## 📖 참고 사항 <!-- 레퍼런스, 스크린샷 등을 넣어 주세요. -->
non_process
test feat 테스트 💎 개발할 기능 대회에 참가한 플레이어에 대한 reject 테스트 관전자 플레이어 조회 테스트 관리자 부여 테스트 📖 참고 사항
0
43,083
7,021,155,118
IssuesEvent
2017-12-22 02:04:54
stan-dev/math
https://api.github.com/repos/stan-dev/math
closed
bug in neg_binomial_2_log_glm_lpmf.hpp is breaking stan/math
documentation
#### Summary: Trying to run a simple hello_world example from the Readme on a fresh clone of the `develop` branch leads to errors and failure of the program. The errors originate from the file `neg_binomila_2_log_glm_lpmf.hpp`. By commenting out line 292 in `~/stan-dev/math/stan/math/prim/mat.hpp` like so `//#include <stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp>` , the program is able to run as expected. #### Description: When trying to run the hello world example from the Readme of `stan-dev/math` (printing the output of `normal_log(1,2,3)`, erros are raised by `neg_binomila_2_log_glm_lpmf.hpp` and execution fails. Here is a print out of the error logs: ``` | ~/github_projects/stan-dev/math/hello_world @ (develop) | => clang++ -I ~/github_projects/stan-dev/math -I ~/github_projects/stan-dev/math/lib/eigen_3.3.3/ -I ~/github_projects/stan-dev/math/lib/boost_1.64.0/ -I ~/github_projects/stan-dev/math/lib/cvodes_2.9.0/include hello_world.cpp In file included from hello_world.cpp:1: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math.hpp:4: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/rev/mat.hpp:12: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat.hpp:249: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/bernoulli_logit_glm_lpmf.hpp:15: /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/scal/fun/size_zero.hpp:30:31: warning: variadic templates are a C++11 extension [-Wc++11-extensions] template <typename T, typename... Ts> ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/scal/fun/size_zero.hpp:31:31: warning: rvalue references are a C++11 extension [-Wc++11-extensions] inline bool size_zero(T& x, Ts&&... xs) { ^ In file included from hello_world.cpp:1: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math.hpp:4: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/rev/mat.hpp:12: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat.hpp:292: /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:116:39: error: expected expression = theta_dbl.binaryExpr(log_phi, [](const T_partials_return& xx, ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:126:21: error: expected expression [](const T_partials_return& xx) { return lgamma(xx); }) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:132:18: error: expected expression [](const T_partials_return& xx, const T_partials_return& yy) { ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:136:20: error: expected expression [](const T_partials_return& xx) { return lgamma(xx); })) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:146:21: error: expected expression [](const T_partials_return& xx) { return lgamma(xx); }) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:178:18: error: expected expression [](const T_partials_return& xx) { return digamma(xx); }) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:180:18: error: expected expression [](const T_partials_return& xx) { return digamma(xx); })) ^ 2 warnings and 7 errors generated. ``` #### Reproducible Steps: 1. Clone `develop` branch 2. create hello_world folder and hello_world.cpp file with the content ``` #include <stan/math.hpp> #include <iostream> int main() { std::cout << "log normal(1 | 2, 3)=" << stan::math::normal_log(1, 2, 3) << std::endl; } ``` 3. in the shell run the clang++ command with the arguments as decribed in the Readme. You should get the errors printed above. 4. Comment out line 292 in `~/stan-dev/math/stan/math/prim/mat.hpp` and run step 3 again. The program should succeed. #### Current Version: develop
1.0
bug in neg_binomial_2_log_glm_lpmf.hpp is breaking stan/math - #### Summary: Trying to run a simple hello_world example from the Readme on a fresh clone of the `develop` branch leads to errors and failure of the program. The errors originate from the file `neg_binomila_2_log_glm_lpmf.hpp`. By commenting out line 292 in `~/stan-dev/math/stan/math/prim/mat.hpp` like so `//#include <stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp>` , the program is able to run as expected. #### Description: When trying to run the hello world example from the Readme of `stan-dev/math` (printing the output of `normal_log(1,2,3)`, erros are raised by `neg_binomila_2_log_glm_lpmf.hpp` and execution fails. Here is a print out of the error logs: ``` | ~/github_projects/stan-dev/math/hello_world @ (develop) | => clang++ -I ~/github_projects/stan-dev/math -I ~/github_projects/stan-dev/math/lib/eigen_3.3.3/ -I ~/github_projects/stan-dev/math/lib/boost_1.64.0/ -I ~/github_projects/stan-dev/math/lib/cvodes_2.9.0/include hello_world.cpp In file included from hello_world.cpp:1: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math.hpp:4: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/rev/mat.hpp:12: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat.hpp:249: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/bernoulli_logit_glm_lpmf.hpp:15: /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/scal/fun/size_zero.hpp:30:31: warning: variadic templates are a C++11 extension [-Wc++11-extensions] template <typename T, typename... Ts> ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/scal/fun/size_zero.hpp:31:31: warning: rvalue references are a C++11 extension [-Wc++11-extensions] inline bool size_zero(T& x, Ts&&... xs) { ^ In file included from hello_world.cpp:1: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math.hpp:4: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/rev/mat.hpp:12: In file included from /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat.hpp:292: /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:116:39: error: expected expression = theta_dbl.binaryExpr(log_phi, [](const T_partials_return& xx, ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:126:21: error: expected expression [](const T_partials_return& xx) { return lgamma(xx); }) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:132:18: error: expected expression [](const T_partials_return& xx, const T_partials_return& yy) { ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:136:20: error: expected expression [](const T_partials_return& xx) { return lgamma(xx); })) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:146:21: error: expected expression [](const T_partials_return& xx) { return lgamma(xx); }) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:178:18: error: expected expression [](const T_partials_return& xx) { return digamma(xx); }) ^ /Users/youcefmsaid/github_projects/stan-dev/math/stan/math/prim/mat/prob/neg_binomial_2_log_glm_lpmf.hpp:180:18: error: expected expression [](const T_partials_return& xx) { return digamma(xx); })) ^ 2 warnings and 7 errors generated. ``` #### Reproducible Steps: 1. Clone `develop` branch 2. create hello_world folder and hello_world.cpp file with the content ``` #include <stan/math.hpp> #include <iostream> int main() { std::cout << "log normal(1 | 2, 3)=" << stan::math::normal_log(1, 2, 3) << std::endl; } ``` 3. in the shell run the clang++ command with the arguments as decribed in the Readme. You should get the errors printed above. 4. Comment out line 292 in `~/stan-dev/math/stan/math/prim/mat.hpp` and run step 3 again. The program should succeed. #### Current Version: develop
non_process
bug in neg binomial log glm lpmf hpp is breaking stan math summary trying to run a simple hello world example from the readme on a fresh clone of the develop branch leads to errors and failure of the program the errors originate from the file neg binomila log glm lpmf hpp by commenting out line in stan dev math stan math prim mat hpp like so include the program is able to run as expected description when trying to run the hello world example from the readme of stan dev math printing the output of normal log erros are raised by neg binomila log glm lpmf hpp and execution fails here is a print out of the error logs github projects stan dev math hello world develop clang i github projects stan dev math i github projects stan dev math lib eigen i github projects stan dev math lib boost i github projects stan dev math lib cvodes include hello world cpp in file included from hello world cpp in file included from users youcefmsaid github projects stan dev math stan math hpp in file included from users youcefmsaid github projects stan dev math stan math rev mat hpp in file included from users youcefmsaid github projects stan dev math stan math prim mat hpp in file included from users youcefmsaid github projects stan dev math stan math prim mat prob bernoulli logit glm lpmf hpp users youcefmsaid github projects stan dev math stan math prim scal fun size zero hpp warning variadic templates are a c extension template users youcefmsaid github projects stan dev math stan math prim scal fun size zero hpp warning rvalue references are a c extension inline bool size zero t x ts xs in file included from hello world cpp in file included from users youcefmsaid github projects stan dev math stan math hpp in file included from users youcefmsaid github projects stan dev math stan math rev mat hpp in file included from users youcefmsaid github projects stan dev math stan math prim mat hpp users youcefmsaid github projects stan dev math stan math prim mat prob neg binomial log glm lpmf hpp error expected expression theta dbl binaryexpr log phi const t partials return xx users youcefmsaid github projects stan dev math stan math prim mat prob neg binomial log glm lpmf hpp error expected expression const t partials return xx return lgamma xx users youcefmsaid github projects stan dev math stan math prim mat prob neg binomial log glm lpmf hpp error expected expression const t partials return xx const t partials return yy users youcefmsaid github projects stan dev math stan math prim mat prob neg binomial log glm lpmf hpp error expected expression const t partials return xx return lgamma xx users youcefmsaid github projects stan dev math stan math prim mat prob neg binomial log glm lpmf hpp error expected expression const t partials return xx return lgamma xx users youcefmsaid github projects stan dev math stan math prim mat prob neg binomial log glm lpmf hpp error expected expression const t partials return xx return digamma xx users youcefmsaid github projects stan dev math stan math prim mat prob neg binomial log glm lpmf hpp error expected expression const t partials return xx return digamma xx warnings and errors generated reproducible steps clone develop branch create hello world folder and hello world cpp file with the content include include int main std cout log normal stan math normal log std endl in the shell run the clang command with the arguments as decribed in the readme you should get the errors printed above comment out line in stan dev math stan math prim mat hpp and run step again the program should succeed current version develop
0
103,847
11,383,160,099
IssuesEvent
2020-01-29 04:55:35
Seneca-CDOT/telescope
https://api.github.com/repos/Seneca-CDOT/telescope
closed
Documentation for GraphQL
area: back-end area: web server has-PR type: documentation
**What would you like to be added**: We need documentation for our [GraphQL](https://github.com/Seneca-CDOT/telescope/blob/master/src/backend/web/graphql/index.js) endpoint containing information about how it works and the types and queries that it uses. Including a few examples would be very helpful. **Why would you like this to be added**: This will make it easier for backend developers to add, remove or modify queries, and for frontend developers to take full advantage of Graphql's features (single API call, tailored requests, no over-under fetching, faster prototyping, etc).
1.0
Documentation for GraphQL - **What would you like to be added**: We need documentation for our [GraphQL](https://github.com/Seneca-CDOT/telescope/blob/master/src/backend/web/graphql/index.js) endpoint containing information about how it works and the types and queries that it uses. Including a few examples would be very helpful. **Why would you like this to be added**: This will make it easier for backend developers to add, remove or modify queries, and for frontend developers to take full advantage of Graphql's features (single API call, tailored requests, no over-under fetching, faster prototyping, etc).
non_process
documentation for graphql what would you like to be added we need documentation for our endpoint containing information about how it works and the types and queries that it uses including a few examples would be very helpful why would you like this to be added this will make it easier for backend developers to add remove or modify queries and for frontend developers to take full advantage of graphql s features single api call tailored requests no over under fetching faster prototyping etc
0
2,743
5,637,413,310
IssuesEvent
2017-04-06 09:06:17
our-city-app/oca-backend
https://api.github.com/repos/our-city-app/oca-backend
closed
channel not reconnecting?
priority_critical process_wontfix type_bug
Seen this at the development environment and on the tablets of the sales ppl. When a service is being created a progress bar is shown, but never goes away. It looks like the channel gets disconnected, but doesnt automatically reconnect.
1.0
channel not reconnecting? - Seen this at the development environment and on the tablets of the sales ppl. When a service is being created a progress bar is shown, but never goes away. It looks like the channel gets disconnected, but doesnt automatically reconnect.
process
channel not reconnecting seen this at the development environment and on the tablets of the sales ppl when a service is being created a progress bar is shown but never goes away it looks like the channel gets disconnected but doesnt automatically reconnect
1
289,731
8,875,786,805
IssuesEvent
2019-01-12 07:58:28
Stivius/XiboLinuxStack
https://api.github.com/repos/Stivius/XiboLinuxStack
closed
Layout replacement
medium priority task
The issue #155 is also related to this one - [x] Adapt event system (add event queue, refactor) - [x] Check all possible leaks
1.0
Layout replacement - The issue #155 is also related to this one - [x] Adapt event system (add event queue, refactor) - [x] Check all possible leaks
non_process
layout replacement the issue is also related to this one adapt event system add event queue refactor check all possible leaks
0
51,825
7,733,914,247
IssuesEvent
2018-05-26 17:31:35
pmd/pmd
https://api.github.com/repos/pmd/pmd
closed
[core] JCommander's help text for option -min is wrong
has:pr in:documentation
**Affects PMD version:** 6.3.0 The `-min` CLI option (rule priority threshold) takes an integer argument between 1 and 5. However, JCommander assumes it's an enum and displays the following help: ``` -minimumpriority, -min Rule priority threshold; rules with lower priority than configured here won't be used. Default is '5' which is the lowest priority. Default: Low Possible Values: [High, Medium High, Medium, Medium Low, Low] ``` Using the name of an enum constant makes the parameter parsing fail: https://github.com/pmd/pmd/blob/master/pmd-core/src/main/java/net/sourceforge/pmd/cli/PMDParameters.java#L132 Should we allow mentioning the names of the constants? Or just fix the JCommander help? For some reason, only `-min 4` and `-min 5` work as expected (at least on my machine). Using eg `-min 3` makes PMD fail with ``` GRAVE: No rules found. Maybe you mispelled a rule name? (/home/clifrr/Bureau/import_sample/rset.xml) ```
1.0
[core] JCommander's help text for option -min is wrong - **Affects PMD version:** 6.3.0 The `-min` CLI option (rule priority threshold) takes an integer argument between 1 and 5. However, JCommander assumes it's an enum and displays the following help: ``` -minimumpriority, -min Rule priority threshold; rules with lower priority than configured here won't be used. Default is '5' which is the lowest priority. Default: Low Possible Values: [High, Medium High, Medium, Medium Low, Low] ``` Using the name of an enum constant makes the parameter parsing fail: https://github.com/pmd/pmd/blob/master/pmd-core/src/main/java/net/sourceforge/pmd/cli/PMDParameters.java#L132 Should we allow mentioning the names of the constants? Or just fix the JCommander help? For some reason, only `-min 4` and `-min 5` work as expected (at least on my machine). Using eg `-min 3` makes PMD fail with ``` GRAVE: No rules found. Maybe you mispelled a rule name? (/home/clifrr/Bureau/import_sample/rset.xml) ```
non_process
jcommander s help text for option min is wrong affects pmd version the min cli option rule priority threshold takes an integer argument between and however jcommander assumes it s an enum and displays the following help minimumpriority min rule priority threshold rules with lower priority than configured here won t be used default is which is the lowest priority default low possible values using the name of an enum constant makes the parameter parsing fail should we allow mentioning the names of the constants or just fix the jcommander help for some reason only min and min work as expected at least on my machine using eg min makes pmd fail with grave no rules found maybe you mispelled a rule name home clifrr bureau import sample rset xml
0
100,741
8,753,273,688
IssuesEvent
2018-12-14 07:45:26
legion-platform/legion
https://api.github.com/repos/legion-platform/legion
closed
Create CI build pipeline for cluster infrastructure components to produce set of deliverable infrastructure units
0.10.0 CI/CD/tests
Create Jenkins build pipeline for cluster infrastructure components. This pipeline should include: 1) build of 3d party customised docker containers. - move infra related dockerfiles (kube-fluentd, elb-security, ?) to k8s/k8s-infra - create jenkins job for Infra images build - update release pipeline with new step for infra - update playbooks\charts\etc with new images location - update existing jenkins jobs with new infra version tag parameter - include dex build into this build pipeline. We customised dex in https://github.com/legion-platform/dex/commits/feat/oidc-groups. Rename branch oidc-groups - > legion and set legion release tag. 2) package helm charts (to be discussed ) - public repo for releases https://github.com/helm/charts - private repo for snapshots https://github.com/sonatype-nexus-community/nexus-repository-helm 3) package ansible scripts docker image - public repo for releases docker hub - private repo for snapshots nexus Single version of Legion infrastructure should be set for all packages and containers listed above. Packages and container should be published to repositories (Nexus or public repos - to be discussed) Create cluster pipeline should depend on these containers and packages but not on Legion git repository.
1.0
Create CI build pipeline for cluster infrastructure components to produce set of deliverable infrastructure units - Create Jenkins build pipeline for cluster infrastructure components. This pipeline should include: 1) build of 3d party customised docker containers. - move infra related dockerfiles (kube-fluentd, elb-security, ?) to k8s/k8s-infra - create jenkins job for Infra images build - update release pipeline with new step for infra - update playbooks\charts\etc with new images location - update existing jenkins jobs with new infra version tag parameter - include dex build into this build pipeline. We customised dex in https://github.com/legion-platform/dex/commits/feat/oidc-groups. Rename branch oidc-groups - > legion and set legion release tag. 2) package helm charts (to be discussed ) - public repo for releases https://github.com/helm/charts - private repo for snapshots https://github.com/sonatype-nexus-community/nexus-repository-helm 3) package ansible scripts docker image - public repo for releases docker hub - private repo for snapshots nexus Single version of Legion infrastructure should be set for all packages and containers listed above. Packages and container should be published to repositories (Nexus or public repos - to be discussed) Create cluster pipeline should depend on these containers and packages but not on Legion git repository.
non_process
create ci build pipeline for cluster infrastructure components to produce set of deliverable infrastructure units create jenkins build pipeline for cluster infrastructure components this pipeline should include build of party customised docker containers move infra related dockerfiles kube fluentd elb security to infra create jenkins job for infra images build update release pipeline with new step for infra update playbooks charts etc with new images location update existing jenkins jobs with new infra version tag parameter include dex build into this build pipeline we customised dex in rename branch oidc groups legion and set legion release tag package helm charts to be discussed public repo for releases private repo for snapshots package ansible scripts docker image public repo for releases docker hub private repo for snapshots nexus single version of legion infrastructure should be set for all packages and containers listed above packages and container should be published to repositories nexus or public repos to be discussed create cluster pipeline should depend on these containers and packages but not on legion git repository
0
96,531
10,934,659,921
IssuesEvent
2019-11-24 13:18:18
Varying-Vagrant-Vagrants/VVV
https://api.github.com/repos/Varying-Vagrant-Vagrants/VVV
reopened
Documentation about config needs updating
documentation good first issue
Per the release notes for 3.2.0: "The biggest change is that we no longer use vvv-custom.yml, we’ve moved to config/config.yml." The new site setup documentation here: _https://varyingvagrantvagrants.org/docs/en-US/adding-a-new-site/_ still makes reference to the old `vvv-custom.yml` file setup, which is more than a bit confusing. Sooo I feel like I now need some clarification: in the old setup (`vvv-config.yml` + `vvv-custom.yml`), the ostensible purpose of vvv-custom.yml was to prevent it from inadvertently being overwritten by pulling a VVV update, right? In the new setup, however, there's only the one `config.yml` — no custom. Is it safe to assume that the intention is that this file (derived from `default-config.yml`) will also not be overwritten in the future? Can we treat `config.yml` as functionally equivalent to `vvv-custom.yml`?
1.0
Documentation about config needs updating - Per the release notes for 3.2.0: "The biggest change is that we no longer use vvv-custom.yml, we’ve moved to config/config.yml." The new site setup documentation here: _https://varyingvagrantvagrants.org/docs/en-US/adding-a-new-site/_ still makes reference to the old `vvv-custom.yml` file setup, which is more than a bit confusing. Sooo I feel like I now need some clarification: in the old setup (`vvv-config.yml` + `vvv-custom.yml`), the ostensible purpose of vvv-custom.yml was to prevent it from inadvertently being overwritten by pulling a VVV update, right? In the new setup, however, there's only the one `config.yml` — no custom. Is it safe to assume that the intention is that this file (derived from `default-config.yml`) will also not be overwritten in the future? Can we treat `config.yml` as functionally equivalent to `vvv-custom.yml`?
non_process
documentation about config needs updating per the release notes for the biggest change is that we no longer use vvv custom yml we’ve moved to config config yml the new site setup documentation here still makes reference to the old vvv custom yml file setup which is more than a bit confusing sooo i feel like i now need some clarification in the old setup vvv config yml vvv custom yml the ostensible purpose of vvv custom yml was to prevent it from inadvertently being overwritten by pulling a vvv update right in the new setup however there s only the one config yml — no custom is it safe to assume that the intention is that this file derived from default config yml will also not be overwritten in the future can we treat config yml as functionally equivalent to vvv custom yml
0
18,157
24,193,178,049
IssuesEvent
2022-09-23 19:55:04
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Acceptance tests fail due to rate limiting logic
enhancement P2 process
### Problem Acceptance tests fail due to rate limiting logic ### Solution Add check to manage rate limiting check ### Alternatives _No response_
1.0
Acceptance tests fail due to rate limiting logic - ### Problem Acceptance tests fail due to rate limiting logic ### Solution Add check to manage rate limiting check ### Alternatives _No response_
process
acceptance tests fail due to rate limiting logic problem acceptance tests fail due to rate limiting logic solution add check to manage rate limiting check alternatives no response
1
1,572
4,167,290,697
IssuesEvent
2016-06-20 08:57:18
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Permission denied during partitioning disks
priority_minor process_wontfix type_bug
``` root@cmp01:~# parted -m /dev/sda unit B print -s BYT; /dev/sda:90028302336B:scsi:512:512:gpt:ATA KINGSTON SVP200S; ovs@cmp01:/root$ parted -m /dev/sda unit B print -s Error: Error opening /dev/sda: Permission denied ```
1.0
Permission denied during partitioning disks - ``` root@cmp01:~# parted -m /dev/sda unit B print -s BYT; /dev/sda:90028302336B:scsi:512:512:gpt:ATA KINGSTON SVP200S; ovs@cmp01:/root$ parted -m /dev/sda unit B print -s Error: Error opening /dev/sda: Permission denied ```
process
permission denied during partitioning disks root parted m dev sda unit b print s byt dev sda scsi gpt ata kingston ovs root parted m dev sda unit b print s error error opening dev sda permission denied
1
16,919
22,266,202,416
IssuesEvent
2022-06-10 07:42:31
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Add acceptance test support for eth_getTransactionCount
enhancement P2 process
### Problem The current acceptance tests implemented in https://github.com/hashgraph/hedera-json-rpc-relay/pull/119 was not able to include `eth_getTransactionCount` since it kept returning ``` err: { "type": "PrecheckStatusError", "message": "transaction 0.0.2@1654029652.985219400 failed precheck with status INVALID_ACCOUNT_ID", "stack": StatusError: transaction 0.0.2@1654029652.985219400 failed precheck with status INVALID_ACCOUNT_ID at new PrecheckStatusError (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/PrecheckStatusError.cjs:43:5) at AccountInfoQuery._mapStatusError (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/query/Query.cjs:431:12) at CostQuery._mapStatusError (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/query/CostQuery.cjs:155:24) at CostQuery.execute (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/Executable.cjs:519:22) at processTicksAndRejections (node:internal/process/task_queues:96:5) at AccountInfoQuery.getCost (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/account/AccountInfoQuery.cjs:144:16) at AccountInfoQuery._beforeExecute (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/query/Query.cjs:267:28) at AccountInfoQuery.execute (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/Executable.cjs:411:5) "name": "StatusError", "status": { "_code": 15 }, ``` ### Solution Add a test that calls `eth_getTransactionCount ` for the primary and secondary accounts setup ### Alternatives _No response_
1.0
Add acceptance test support for eth_getTransactionCount - ### Problem The current acceptance tests implemented in https://github.com/hashgraph/hedera-json-rpc-relay/pull/119 was not able to include `eth_getTransactionCount` since it kept returning ``` err: { "type": "PrecheckStatusError", "message": "transaction 0.0.2@1654029652.985219400 failed precheck with status INVALID_ACCOUNT_ID", "stack": StatusError: transaction 0.0.2@1654029652.985219400 failed precheck with status INVALID_ACCOUNT_ID at new PrecheckStatusError (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/PrecheckStatusError.cjs:43:5) at AccountInfoQuery._mapStatusError (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/query/Query.cjs:431:12) at CostQuery._mapStatusError (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/query/CostQuery.cjs:155:24) at CostQuery.execute (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/Executable.cjs:519:22) at processTicksAndRejections (node:internal/process/task_queues:96:5) at AccountInfoQuery.getCost (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/account/AccountInfoQuery.cjs:144:16) at AccountInfoQuery._beforeExecute (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/query/Query.cjs:267:28) at AccountInfoQuery.execute (.../hedera-json-rpc-relay/packages/relay/node_modules/@hashgraph/sdk/lib/Executable.cjs:411:5) "name": "StatusError", "status": { "_code": 15 }, ``` ### Solution Add a test that calls `eth_getTransactionCount ` for the primary and secondary accounts setup ### Alternatives _No response_
process
add acceptance test support for eth gettransactioncount problem the current acceptance tests implemented in was not able to include eth gettransactioncount since it kept returning err type precheckstatuserror message transaction failed precheck with status invalid account id stack statuserror transaction failed precheck with status invalid account id at new precheckstatuserror hedera json rpc relay packages relay node modules hashgraph sdk lib precheckstatuserror cjs at accountinfoquery mapstatuserror hedera json rpc relay packages relay node modules hashgraph sdk lib query query cjs at costquery mapstatuserror hedera json rpc relay packages relay node modules hashgraph sdk lib query costquery cjs at costquery execute hedera json rpc relay packages relay node modules hashgraph sdk lib executable cjs at processticksandrejections node internal process task queues at accountinfoquery getcost hedera json rpc relay packages relay node modules hashgraph sdk lib account accountinfoquery cjs at accountinfoquery beforeexecute hedera json rpc relay packages relay node modules hashgraph sdk lib query query cjs at accountinfoquery execute hedera json rpc relay packages relay node modules hashgraph sdk lib executable cjs name statuserror status code solution add a test that calls eth gettransactioncount for the primary and secondary accounts setup alternatives no response
1
8,798
11,908,251,569
IssuesEvent
2020-03-31 00:24:39
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Python error when running a Processing model
Bug Feedback Processing
Author Name: **Magnus Nilsson** (Magnus Nilsson) Original Redmine Issue: [20972](https://issues.qgis.org/issues/20972) Affected QGIS version: 3.4.3 Redmine category:processing/modeller --- I am trying to use an online CSV file and plot its content as points using a Processing model. These steps causes a Python error: 1) Open a new model 2) Add the algorithm for downloading a file. I used https://earthquake.usgs.gov/earthquakes/feed/v1.0/summary/4.5_day.csv. 3) Add the algorithm for creating points from a table and configure it 4) Run the model 5) I now get this error message: Traceback (most recent call last): File "C:/PROGRA~1/GIS/QGIS/apps/qgis/./python/plugins\processing\algs\qgis\PointsLayerFromTable.py", line 91, in processAlgorithm raise QgsProcessingException(self.invalidSourceError(parameters, self.INPUT)) _core.QgsProcessingException: Kunde inte läsa in källager för INPUT: C:/Users/Magnus/AppData/Local/Temp/processing_fa76c51163e14b998c7ffcb4867abf80/980f76d9052042cd8525137ae9ede68f/OUTPUT.file ej hittad Ett fel uppstod vid körning av Skapa punktlager från tabell Ett fel uppstod vid körning av Skapa punktlager från tabell Beräkningen misslyckades efter 0.38 sekunder I expected the data from the download algorithm to be available for other algorithms. Is this incorrect or is this a bug? --- - [CSV from web.model3](https://issues.qgis.org/attachments/download/14083/CSV%20from%20web.model3) (Magnus Nilsson)
1.0
Python error when running a Processing model - Author Name: **Magnus Nilsson** (Magnus Nilsson) Original Redmine Issue: [20972](https://issues.qgis.org/issues/20972) Affected QGIS version: 3.4.3 Redmine category:processing/modeller --- I am trying to use an online CSV file and plot its content as points using a Processing model. These steps causes a Python error: 1) Open a new model 2) Add the algorithm for downloading a file. I used https://earthquake.usgs.gov/earthquakes/feed/v1.0/summary/4.5_day.csv. 3) Add the algorithm for creating points from a table and configure it 4) Run the model 5) I now get this error message: Traceback (most recent call last): File "C:/PROGRA~1/GIS/QGIS/apps/qgis/./python/plugins\processing\algs\qgis\PointsLayerFromTable.py", line 91, in processAlgorithm raise QgsProcessingException(self.invalidSourceError(parameters, self.INPUT)) _core.QgsProcessingException: Kunde inte läsa in källager för INPUT: C:/Users/Magnus/AppData/Local/Temp/processing_fa76c51163e14b998c7ffcb4867abf80/980f76d9052042cd8525137ae9ede68f/OUTPUT.file ej hittad Ett fel uppstod vid körning av Skapa punktlager från tabell Ett fel uppstod vid körning av Skapa punktlager från tabell Beräkningen misslyckades efter 0.38 sekunder I expected the data from the download algorithm to be available for other algorithms. Is this incorrect or is this a bug? --- - [CSV from web.model3](https://issues.qgis.org/attachments/download/14083/CSV%20from%20web.model3) (Magnus Nilsson)
process
python error when running a processing model author name magnus nilsson magnus nilsson original redmine issue affected qgis version redmine category processing modeller i am trying to use an online csv file and plot its content as points using a processing model these steps causes a python error open a new model add the algorithm for downloading a file i used add the algorithm for creating points from a table and configure it run the model i now get this error message traceback most recent call last file c progra gis qgis apps qgis python plugins processing algs qgis pointslayerfromtable py line in processalgorithm raise qgsprocessingexception self invalidsourceerror parameters self input core qgsprocessingexception kunde inte läsa in källager för input c users magnus appdata local temp processing output file ej hittad ett fel uppstod vid körning av skapa punktlager från tabell ett fel uppstod vid körning av skapa punktlager från tabell beräkningen misslyckades efter sekunder i expected the data from the download algorithm to be available for other algorithms is this incorrect or is this a bug magnus nilsson
1
2,201
5,040,736,000
IssuesEvent
2016-12-19 07:23:41
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Redundant sections in Processing?
Processing
Many sections under [Writing new Processing algorithms as python scripts](http://docs.qgis.org/2.8/en/docs/user_manual/processing/scripts.html#writing-new-processing-algorithms-as-python-scripts) chapter have (almost) the same texts than those under [Creating scripts and running them from the toolbox](http://docs.qgis.org/2.8/en/docs/user_manual/processing/console.html#creating-scripts-and-running-them-from-the-toolbox) section and followers. I think they should be merged so that we have only one place to update texts.
1.0
Redundant sections in Processing? - Many sections under [Writing new Processing algorithms as python scripts](http://docs.qgis.org/2.8/en/docs/user_manual/processing/scripts.html#writing-new-processing-algorithms-as-python-scripts) chapter have (almost) the same texts than those under [Creating scripts and running them from the toolbox](http://docs.qgis.org/2.8/en/docs/user_manual/processing/console.html#creating-scripts-and-running-them-from-the-toolbox) section and followers. I think they should be merged so that we have only one place to update texts.
process
redundant sections in processing many sections under chapter have almost the same texts than those under section and followers i think they should be merged so that we have only one place to update texts
1
18,383
24,514,448,608
IssuesEvent
2022-10-11 02:46:14
opensearch-project/data-prepper
https://api.github.com/repos/opensearch-project/data-prepper
opened
Processor to map values to alternate values
enhancement plugin - processor backlog
**Is your feature request related to a problem? Please describe.** Some values, such as codes, can be assigned alternate meaning. I'd like to map one value to another. HTTP status codes are a classic example. I might want to map fields with a value of `200` to `OK` and fields with a value of `404` to `Not Found`. **Describe the solution you'd like** Create a new processor which can map values. It should take the following. * A source key * A destination key * A map of values from the source to values to set in the destination. For example: ``` processor: map_value: source_key: /status_code destination_key: /status map: 200: OK 404: Not Found ``` I'd like a better name than `map_value`. Another interesting possibility would be supporting ranges for numbers. Some examples: ``` processor: map_value: source_key: /status_code destination_key: /status range: 200-399: Success 400-499: Client Error 500-599: Server Error ``` And perhaps comma-delimited values: ``` processor: map_value: source_key: /status_code destination_key: /status range: 200,201: Success 401,403: Authentication 500-599: Server Error ```
1.0
Processor to map values to alternate values - **Is your feature request related to a problem? Please describe.** Some values, such as codes, can be assigned alternate meaning. I'd like to map one value to another. HTTP status codes are a classic example. I might want to map fields with a value of `200` to `OK` and fields with a value of `404` to `Not Found`. **Describe the solution you'd like** Create a new processor which can map values. It should take the following. * A source key * A destination key * A map of values from the source to values to set in the destination. For example: ``` processor: map_value: source_key: /status_code destination_key: /status map: 200: OK 404: Not Found ``` I'd like a better name than `map_value`. Another interesting possibility would be supporting ranges for numbers. Some examples: ``` processor: map_value: source_key: /status_code destination_key: /status range: 200-399: Success 400-499: Client Error 500-599: Server Error ``` And perhaps comma-delimited values: ``` processor: map_value: source_key: /status_code destination_key: /status range: 200,201: Success 401,403: Authentication 500-599: Server Error ```
process
processor to map values to alternate values is your feature request related to a problem please describe some values such as codes can be assigned alternate meaning i d like to map one value to another http status codes are a classic example i might want to map fields with a value of to ok and fields with a value of to not found describe the solution you d like create a new processor which can map values it should take the following a source key a destination key a map of values from the source to values to set in the destination for example processor map value source key status code destination key status map ok not found i d like a better name than map value another interesting possibility would be supporting ranges for numbers some examples processor map value source key status code destination key status range success client error server error and perhaps comma delimited values processor map value source key status code destination key status range success authentication server error
1
4,999
7,834,674,006
IssuesEvent
2018-06-16 17:01:00
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
closed
Incorrect html5 figure breaks css
bug postprocessing schema
Continuing my adventures with the caption package, the following tex ```tex \documentclass{article} \usepackage{caption} \begin{document} \begin{center} \marginpar{\captionsetup{type=figure}\caption{good caption}} \end{center} \begin{center} text \marginpar{\captionsetup{type=figure}\caption{bad caption}} \end{center} \end{document} ``` produces html I would expect, but that turns out to be incorrect. A simplified version of that html is the following: ```html <!doctype html> <html> <head> <meta charset='utf-8'> <title>Test</title> <style type="text/css"> span figure { border: 1px solid green; } body > figure { border: 1px solid red; } </style> </head> <body> <span><figure>good content</figure></span> <p><span><figure>bad content</figure></span></p> </body> </html> ``` The problem is that it seems figure tags aren't allowed in p tags ([a severely out of date source](https://stackoverflow.com/q/5163294/2336725)). Therefore, Chrome and Firefox (at least) treat the second figure as if it were ```html <p><span></span></p> <figure>bad content</figure> <p></p> ``` This means that the default css for margin notes no longer applies. (The [W3C validator](https://validator.w3.org/) also complains about the first span > figure, but Chrome and Firefox at least render it correctly.)
1.0
Incorrect html5 figure breaks css - Continuing my adventures with the caption package, the following tex ```tex \documentclass{article} \usepackage{caption} \begin{document} \begin{center} \marginpar{\captionsetup{type=figure}\caption{good caption}} \end{center} \begin{center} text \marginpar{\captionsetup{type=figure}\caption{bad caption}} \end{center} \end{document} ``` produces html I would expect, but that turns out to be incorrect. A simplified version of that html is the following: ```html <!doctype html> <html> <head> <meta charset='utf-8'> <title>Test</title> <style type="text/css"> span figure { border: 1px solid green; } body > figure { border: 1px solid red; } </style> </head> <body> <span><figure>good content</figure></span> <p><span><figure>bad content</figure></span></p> </body> </html> ``` The problem is that it seems figure tags aren't allowed in p tags ([a severely out of date source](https://stackoverflow.com/q/5163294/2336725)). Therefore, Chrome and Firefox (at least) treat the second figure as if it were ```html <p><span></span></p> <figure>bad content</figure> <p></p> ``` This means that the default css for margin notes no longer applies. (The [W3C validator](https://validator.w3.org/) also complains about the first span > figure, but Chrome and Firefox at least render it correctly.)
process
incorrect figure breaks css continuing my adventures with the caption package the following tex tex documentclass article usepackage caption begin document begin center marginpar captionsetup type figure caption good caption end center begin center text marginpar captionsetup type figure caption bad caption end center end document produces html i would expect but that turns out to be incorrect a simplified version of that html is the following html test span figure border solid green body figure border solid red good content bad content the problem is that it seems figure tags aren t allowed in p tags therefore chrome and firefox at least treat the second figure as if it were html bad content this means that the default css for margin notes no longer applies the also complains about the first span figure but chrome and firefox at least render it correctly
1
18,724
24,611,464,710
IssuesEvent
2022-10-14 22:06:14
GoogleCloudPlatform/cloud-ops-sandbox
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
closed
chore: Modify Github workflows
priority: p1 type: process
Clean up the existing collection of workflows as a part of the migration to the trunk-based branching model. Implement three workflows that should cover the new process: 1. CI workflow to run building, linting and local (a.k.a. unit) testing for all artifacts in the repo: Terraform configurations, Sandbox CLI and supported application configurations. 2. E2e workflow to run end-to-end (a.k.a. integration) testing for `deploy`, `remove` and `recipes` commands of Sandbox CLI using a test configurations. 3. Release workflow to run end-to-end testing for all supported applications before creating a new release. Make workflows to use Github-managed runners instead of self-managed that are in use now. Use more secure [keyless authentication](https://cloud.google.com/blog/products/identity-security/enabling-keyless-authentication-from-github-actions) to authenticate the runners for e2e and release workflows.
1.0
chore: Modify Github workflows - Clean up the existing collection of workflows as a part of the migration to the trunk-based branching model. Implement three workflows that should cover the new process: 1. CI workflow to run building, linting and local (a.k.a. unit) testing for all artifacts in the repo: Terraform configurations, Sandbox CLI and supported application configurations. 2. E2e workflow to run end-to-end (a.k.a. integration) testing for `deploy`, `remove` and `recipes` commands of Sandbox CLI using a test configurations. 3. Release workflow to run end-to-end testing for all supported applications before creating a new release. Make workflows to use Github-managed runners instead of self-managed that are in use now. Use more secure [keyless authentication](https://cloud.google.com/blog/products/identity-security/enabling-keyless-authentication-from-github-actions) to authenticate the runners for e2e and release workflows.
process
chore modify github workflows clean up the existing collection of workflows as a part of the migration to the trunk based branching model implement three workflows that should cover the new process ci workflow to run building linting and local a k a unit testing for all artifacts in the repo terraform configurations sandbox cli and supported application configurations workflow to run end to end a k a integration testing for deploy remove and recipes commands of sandbox cli using a test configurations release workflow to run end to end testing for all supported applications before creating a new release make workflows to use github managed runners instead of self managed that are in use now use more secure to authenticate the runners for and release workflows
1
332,249
29,193,324,883
IssuesEvent
2023-05-19 22:56:41
litentry/litentry-parachain
https://api.github.com/repos/litentry/litentry-parachain
opened
Use consistent types for `WorkerRpcReturnValue` and `WorkerRpcReturnString`
I3-high D5-testing
### Context We have some inconsistent type definitions for these two types, which can cause some problems in decoding as they have incompatible types. https://github.com/litentry/litentry-parachain/blob/d3f6570b256bd5fed0a637bea2d511c4e433bb5c/tee-worker/ts-tests/common/type-definitions.ts#L23-L31 versus https://github.com/litentry/litentry-parachain/blob/d3f6570b256bd5fed0a637bea2d511c4e433bb5c/tee-worker/ts-tests/interfaces/identity/definitions.ts#L3-L10 We should stick to the second definition. --- :heavy_check_mark: Please set appropriate **labels** and **assignees** if applicable.
1.0
Use consistent types for `WorkerRpcReturnValue` and `WorkerRpcReturnString` - ### Context We have some inconsistent type definitions for these two types, which can cause some problems in decoding as they have incompatible types. https://github.com/litentry/litentry-parachain/blob/d3f6570b256bd5fed0a637bea2d511c4e433bb5c/tee-worker/ts-tests/common/type-definitions.ts#L23-L31 versus https://github.com/litentry/litentry-parachain/blob/d3f6570b256bd5fed0a637bea2d511c4e433bb5c/tee-worker/ts-tests/interfaces/identity/definitions.ts#L3-L10 We should stick to the second definition. --- :heavy_check_mark: Please set appropriate **labels** and **assignees** if applicable.
non_process
use consistent types for workerrpcreturnvalue and workerrpcreturnstring context we have some inconsistent type definitions for these two types which can cause some problems in decoding as they have incompatible types versus we should stick to the second definition heavy check mark please set appropriate labels and assignees if applicable
0
300,929
9,213,936,953
IssuesEvent
2019-03-10 16:02:25
TGDF/official-site
https://api.github.com/repos/TGDF/official-site
closed
[navbar & CFP toggle] When CFP toggle is on, Agenda button string on navbar should change
priority
Currently,議程 (Agenda) is the page that can be toggled between Call For Presentation or not. But the name of the button itself can be misleading. The toggle should change the name of that button. CFP on: agenda page's button string should be "講題徵集" in Traditional Chinese; "Call For Presentation" in English (if English string is too long, "CFP" should do) CFP off: agenda page's button string should be "議程表" in Traditional Chinese; "Agenda" in English
1.0
[navbar & CFP toggle] When CFP toggle is on, Agenda button string on navbar should change - Currently,議程 (Agenda) is the page that can be toggled between Call For Presentation or not. But the name of the button itself can be misleading. The toggle should change the name of that button. CFP on: agenda page's button string should be "講題徵集" in Traditional Chinese; "Call For Presentation" in English (if English string is too long, "CFP" should do) CFP off: agenda page's button string should be "議程表" in Traditional Chinese; "Agenda" in English
non_process
when cfp toggle is on agenda button string on navbar should change currently,議程 agenda is the page that can be toggled between call for presentation or not but the name of the button itself can be misleading the toggle should change the name of that button cfp on agenda page s button string should be 講題徵集 in traditional chinese call for presentation in english if english string is too long cfp should do cfp off agenda page s button string should be 議程表 in traditional chinese agenda in english
0
6,567
7,693,054,604
IssuesEvent
2018-05-18 01:11:26
Microsoft/vscode-cpptools
https://api.github.com/repos/Microsoft/vscode-cpptools
closed
Recursive includes can cause IntelliSense to be a lot slower
Language Service bug fixed (release pending) performance
With 0.17.0: If your recursive include expands to too many folders, it can cause the red flame (IntelliSense updating) to take too long (e.g. 15 times more include paths can lead to a 5 times slow down). We've disabled recursive includes "by default" for 0.17.0. We have a fix internally, but we're still working on fixing the scenario when new `#includes` are added.
1.0
Recursive includes can cause IntelliSense to be a lot slower - With 0.17.0: If your recursive include expands to too many folders, it can cause the red flame (IntelliSense updating) to take too long (e.g. 15 times more include paths can lead to a 5 times slow down). We've disabled recursive includes "by default" for 0.17.0. We have a fix internally, but we're still working on fixing the scenario when new `#includes` are added.
non_process
recursive includes can cause intellisense to be a lot slower with if your recursive include expands to too many folders it can cause the red flame intellisense updating to take too long e g times more include paths can lead to a times slow down we ve disabled recursive includes by default for we have a fix internally but we re still working on fixing the scenario when new includes are added
0
18,191
24,239,959,920
IssuesEvent
2022-09-27 05:26:58
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Support split string into arrays for a log attribute
enhancement good first issue priority:p2 processor/transform
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] For example, I have a log line like this ``` {"msg": "xxx", "flag": "A|B|C"} ``` with filelog receiver and its json_parser, I am able to access `flag` field, and I would like to split `A|B|C` by `|` so it will be an array, possible use cases are 1) filter the array elements by other processors 2) query if array contains certain elements at storage side if the storage supports array. **Describe the solution you'd like** Add string split operator in filelog receiver. **Describe alternatives you've considered** Add support in transform processor. **Additional context**
1.0
Support split string into arrays for a log attribute - **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] For example, I have a log line like this ``` {"msg": "xxx", "flag": "A|B|C"} ``` with filelog receiver and its json_parser, I am able to access `flag` field, and I would like to split `A|B|C` by `|` so it will be an array, possible use cases are 1) filter the array elements by other processors 2) query if array contains certain elements at storage side if the storage supports array. **Describe the solution you'd like** Add string split operator in filelog receiver. **Describe alternatives you've considered** Add support in transform processor. **Additional context**
process
support split string into arrays for a log attribute is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when for example i have a log line like this msg xxx flag a b c with filelog receiver and its json parser i am able to access flag field and i would like to split a b c by so it will be an array possible use cases are filter the array elements by other processors query if array contains certain elements at storage side if the storage supports array describe the solution you d like add string split operator in filelog receiver describe alternatives you ve considered add support in transform processor additional context
1
638
3,093,546,755
IssuesEvent
2015-08-27 00:02:59
meteor/meteor
https://api.github.com/repos/meteor/meteor
closed
Preview of batch plugins and other 1.2 features
Project:Release Process
This issue is a good place to track concerns found with the `PLUGINS-PREVIEW@1` release. We've been working hard over the past few months on a number of features for Meteor 1.2. @Slava and I have been focused on adding new capabilities to build plugins, and we have a preview release ready! Try it out with `meteor --release PLUGINS-PREVIEW@1`. This release adds new plugin APIs: `Plugin.registerCompiler`, `Plugin.registerLinter`, and `Plugin.registerMinifier`. `registerCompiler` is a replacement for the now-deprecated `registerSourceHandler` feature. Compilers are like source handlers, but they always run when you're building your app instead of when packages are published, and they have access to all of the relevant files in your app and its packages at once. This lets us implement things like CSS preprocessor `@import`s that work across package boundaries. `registerLinter` allows you to write linters that handle different file types and show warnings when you run your app or publish a package (or run the new `meteor lint` command). `registerMinifier` allows you to define your own minifiers to use instead of the standard minifiers, which are no longer baked into the tool but now can be replaced. The new APIs are documented at https://github.com/meteor/meteor/wiki/Build-Plugins-API Some notes: * You'll need to add the `standard-minifiers` package to your app when testing this. (When this is released for real as Meteor 1.2, `meteor update` will add that to your app automatically.) * The `less` and `stylus` packages now support cross-package imports! If you specify a file like `@import "{package}/file.js"` (with the curly braces), it will load the file from that package. Use `@import "{}/file.js"` to load a file from your app. * There's a new core `jshint` package which is a linter for js files using JSHint. * We've made one backwards-incompatible change to packages. To include static assets in packages, you now need to explicitly specify `{isAsset: true}` in your `api.addFiles` call. This is not necessary for static assets in apps, and doesn't affect published packages, just package sources. * Because of this change to these packages, we've bumped their major version numbers. Packages that include `less` files will need to publish a new version using this preview (or an RC) before you can use them. * We've improved rebuild time in `meteor run` in various ways, including via new caches and new APIs that allow plugins to define their own caches. * There's a lot of other stuff in here too! For example, there's a new core `ecmascript` package which uses Babel to let you write Meteor code in ES2015 (similar to `grigio:babel`). There's automatic compression on the wire for DDP. There's some improvements to Livequery oplog tailing performance. There are several months worth of incremental bugfixes. Pretty exciting! Were not quite ready for the formal 1.2 release candidate process, but we'd love to hear how this release works for your apps! And plugin authors (or those interested in becoming plugin authors), we'd love to see how the new APIs work for you! **Update july 27 (10am PST, by Slava)**: The new `PLUGINS-PREVIEW@2` pre-release has been pushed. The updates from the previous preview: - fixed some importing from imports bugs in Stylus - QA'd and tested on Windows (fixed bugs relevant to handling paths) - all build plugins should prefer using `Plugin.fs` and `Plugin.path` that work with the standard paths (`/C/some path/file.exe` as opposed to `C:\some path\files.exe`) - fixed import in `accounts-ui`'s less file - a `jscript` fix for the `babel-compiler` package - renewed dev-bundle working well for Windows **Issues to resolve**: - [ ] https://github.com/meteor/meteor/issues/4839
1.0
Preview of batch plugins and other 1.2 features - This issue is a good place to track concerns found with the `PLUGINS-PREVIEW@1` release. We've been working hard over the past few months on a number of features for Meteor 1.2. @Slava and I have been focused on adding new capabilities to build plugins, and we have a preview release ready! Try it out with `meteor --release PLUGINS-PREVIEW@1`. This release adds new plugin APIs: `Plugin.registerCompiler`, `Plugin.registerLinter`, and `Plugin.registerMinifier`. `registerCompiler` is a replacement for the now-deprecated `registerSourceHandler` feature. Compilers are like source handlers, but they always run when you're building your app instead of when packages are published, and they have access to all of the relevant files in your app and its packages at once. This lets us implement things like CSS preprocessor `@import`s that work across package boundaries. `registerLinter` allows you to write linters that handle different file types and show warnings when you run your app or publish a package (or run the new `meteor lint` command). `registerMinifier` allows you to define your own minifiers to use instead of the standard minifiers, which are no longer baked into the tool but now can be replaced. The new APIs are documented at https://github.com/meteor/meteor/wiki/Build-Plugins-API Some notes: * You'll need to add the `standard-minifiers` package to your app when testing this. (When this is released for real as Meteor 1.2, `meteor update` will add that to your app automatically.) * The `less` and `stylus` packages now support cross-package imports! If you specify a file like `@import "{package}/file.js"` (with the curly braces), it will load the file from that package. Use `@import "{}/file.js"` to load a file from your app. * There's a new core `jshint` package which is a linter for js files using JSHint. * We've made one backwards-incompatible change to packages. To include static assets in packages, you now need to explicitly specify `{isAsset: true}` in your `api.addFiles` call. This is not necessary for static assets in apps, and doesn't affect published packages, just package sources. * Because of this change to these packages, we've bumped their major version numbers. Packages that include `less` files will need to publish a new version using this preview (or an RC) before you can use them. * We've improved rebuild time in `meteor run` in various ways, including via new caches and new APIs that allow plugins to define their own caches. * There's a lot of other stuff in here too! For example, there's a new core `ecmascript` package which uses Babel to let you write Meteor code in ES2015 (similar to `grigio:babel`). There's automatic compression on the wire for DDP. There's some improvements to Livequery oplog tailing performance. There are several months worth of incremental bugfixes. Pretty exciting! Were not quite ready for the formal 1.2 release candidate process, but we'd love to hear how this release works for your apps! And plugin authors (or those interested in becoming plugin authors), we'd love to see how the new APIs work for you! **Update july 27 (10am PST, by Slava)**: The new `PLUGINS-PREVIEW@2` pre-release has been pushed. The updates from the previous preview: - fixed some importing from imports bugs in Stylus - QA'd and tested on Windows (fixed bugs relevant to handling paths) - all build plugins should prefer using `Plugin.fs` and `Plugin.path` that work with the standard paths (`/C/some path/file.exe` as opposed to `C:\some path\files.exe`) - fixed import in `accounts-ui`'s less file - a `jscript` fix for the `babel-compiler` package - renewed dev-bundle working well for Windows **Issues to resolve**: - [ ] https://github.com/meteor/meteor/issues/4839
process
preview of batch plugins and other features this issue is a good place to track concerns found with the plugins preview release we ve been working hard over the past few months on a number of features for meteor slava and i have been focused on adding new capabilities to build plugins and we have a preview release ready try it out with meteor release plugins preview this release adds new plugin apis plugin registercompiler plugin registerlinter and plugin registerminifier registercompiler is a replacement for the now deprecated registersourcehandler feature compilers are like source handlers but they always run when you re building your app instead of when packages are published and they have access to all of the relevant files in your app and its packages at once this lets us implement things like css preprocessor import s that work across package boundaries registerlinter allows you to write linters that handle different file types and show warnings when you run your app or publish a package or run the new meteor lint command registerminifier allows you to define your own minifiers to use instead of the standard minifiers which are no longer baked into the tool but now can be replaced the new apis are documented at some notes you ll need to add the standard minifiers package to your app when testing this when this is released for real as meteor meteor update will add that to your app automatically the less and stylus packages now support cross package imports if you specify a file like import package file js with the curly braces it will load the file from that package use import file js to load a file from your app there s a new core jshint package which is a linter for js files using jshint we ve made one backwards incompatible change to packages to include static assets in packages you now need to explicitly specify isasset true in your api addfiles call this is not necessary for static assets in apps and doesn t affect published packages just package sources because of this change to these packages we ve bumped their major version numbers packages that include less files will need to publish a new version using this preview or an rc before you can use them we ve improved rebuild time in meteor run in various ways including via new caches and new apis that allow plugins to define their own caches there s a lot of other stuff in here too for example there s a new core ecmascript package which uses babel to let you write meteor code in similar to grigio babel there s automatic compression on the wire for ddp there s some improvements to livequery oplog tailing performance there are several months worth of incremental bugfixes pretty exciting were not quite ready for the formal release candidate process but we d love to hear how this release works for your apps and plugin authors or those interested in becoming plugin authors we d love to see how the new apis work for you update july pst by slava the new plugins preview pre release has been pushed the updates from the previous preview fixed some importing from imports bugs in stylus qa d and tested on windows fixed bugs relevant to handling paths all build plugins should prefer using plugin fs and plugin path that work with the standard paths c some path file exe as opposed to c some path files exe fixed import in accounts ui s less file a jscript fix for the babel compiler package renewed dev bundle working well for windows issues to resolve
1
15,838
20,027,131,538
IssuesEvent
2022-02-01 22:47:44
googleapis/java-vision
https://api.github.com/repos/googleapis/java-vision
closed
release is mis-tagged
type: process api: vision
Relased as [v1.100.7 on sonatype](https://oss.sonatype.org/service/local/repositories/releases/content/com/google/cloud/google-cloud-vision/1.100.7/google-cloud-vision-1.100.7.pom). Fusion build [log](https://g3c.corp.google.com/results/invocations/35659bc5-7120-4b39-ac26-93b4eec7b346/targets/cloud-devrel%2Fclient-libraries%2Fjava%2Fjava-vision%2Frelease%2Fstage/log). However, repo is tagging it as [v1.110.7](https://github.com/googleapis/java-vision/releases/tag/v1.110.7) Discovered through https://github.com/googleapis/java-cloud-bom/issues/735 --> google-cloud-bom is updating to v1.100.7
1.0
release is mis-tagged - Relased as [v1.100.7 on sonatype](https://oss.sonatype.org/service/local/repositories/releases/content/com/google/cloud/google-cloud-vision/1.100.7/google-cloud-vision-1.100.7.pom). Fusion build [log](https://g3c.corp.google.com/results/invocations/35659bc5-7120-4b39-ac26-93b4eec7b346/targets/cloud-devrel%2Fclient-libraries%2Fjava%2Fjava-vision%2Frelease%2Fstage/log). However, repo is tagging it as [v1.110.7](https://github.com/googleapis/java-vision/releases/tag/v1.110.7) Discovered through https://github.com/googleapis/java-cloud-bom/issues/735 --> google-cloud-bom is updating to v1.100.7
process
release is mis tagged relased as fusion build however repo is tagging it as discovered through google cloud bom is updating to
1
123,928
16,551,466,631
IssuesEvent
2021-05-28 09:04:11
google/web-stories-wp
https://api.github.com/repos/google/web-stories-wp
closed
Design System: Add long form snackbar message
Group: Design System Pod: Pea Type: Enhancement
## Feature Description Create a new snackbar version for longer messages where the following is true: Cta should still be docked to the right and vertically centered. See #4957 for more context. ## Alternatives Considered <!-- A clear and concise description of any alternative solutions or features you've considered. --> ## Additional Context https://www.figma.com/file/bMhG3KyrJF8vIAODgmbeqT/Design-System?node-id=3398%3A83619 --- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief <!-- One or more bullet points for how to technically implement the feature. -->
1.0
Design System: Add long form snackbar message - ## Feature Description Create a new snackbar version for longer messages where the following is true: Cta should still be docked to the right and vertically centered. See #4957 for more context. ## Alternatives Considered <!-- A clear and concise description of any alternative solutions or features you've considered. --> ## Additional Context https://www.figma.com/file/bMhG3KyrJF8vIAODgmbeqT/Design-System?node-id=3398%3A83619 --- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief <!-- One or more bullet points for how to technically implement the feature. -->
non_process
design system add long form snackbar message feature description create a new snackbar version for longer messages where the following is true cta should still be docked to the right and vertically centered see for more context alternatives considered additional context do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief
0
312,056
26,833,929,743
IssuesEvent
2023-02-02 17:56:52
PalisadoesFoundation/talawa-api
https://api.github.com/repos/PalisadoesFoundation/talawa-api
closed
Test: src/lib/resolvers/Subscription/directMessageChat.ts
good first issue points 01 test
- Please coordinate **issue assignment** and **PR reviews** with the contributors listed in this issue https://github.com/PalisadoesFoundation/talawa/issues/359 The Talawa-API code base needs to be 100% reliable. This means we need to have 100% test code coverage. - Tests need to be written for file `src/lib/resolvers/Subscription/directMessageChat.ts ` - We will need the API to be refactored for all methods, classes and/or functions found in this file for testing to be correctly executed. - When complete, all all methods, classes and/or functions in the refactored file will need to be tested. These tests must be placed in a single file with the name `__tests__/resolvers/Subscription/directMessageChat.spec.ts`. You may need to create the appropriate directory structure to do this. ### IMPORTANT: Please refer to the parent issue on how to implement these tests correctly: - https://github.com/PalisadoesFoundation/talawa-api/issues/490 ### PR Acceptance Criteria - When complete this file must show **100%** coverage when merged into the code base. This will be clearly visible when you submit your PR. - [The current code coverage for the file can be found by visting this page](https://app.codecov.io/gh/PalisadoesFoundation/talawa-api?displayType=list). Login using your GitHub credentials. - Create your own `codecov.io` to help with testing. - The PR you create will show a report for the code coverage for the file you have added. You can also use that as a guide.
1.0
Test: src/lib/resolvers/Subscription/directMessageChat.ts - - Please coordinate **issue assignment** and **PR reviews** with the contributors listed in this issue https://github.com/PalisadoesFoundation/talawa/issues/359 The Talawa-API code base needs to be 100% reliable. This means we need to have 100% test code coverage. - Tests need to be written for file `src/lib/resolvers/Subscription/directMessageChat.ts ` - We will need the API to be refactored for all methods, classes and/or functions found in this file for testing to be correctly executed. - When complete, all all methods, classes and/or functions in the refactored file will need to be tested. These tests must be placed in a single file with the name `__tests__/resolvers/Subscription/directMessageChat.spec.ts`. You may need to create the appropriate directory structure to do this. ### IMPORTANT: Please refer to the parent issue on how to implement these tests correctly: - https://github.com/PalisadoesFoundation/talawa-api/issues/490 ### PR Acceptance Criteria - When complete this file must show **100%** coverage when merged into the code base. This will be clearly visible when you submit your PR. - [The current code coverage for the file can be found by visting this page](https://app.codecov.io/gh/PalisadoesFoundation/talawa-api?displayType=list). Login using your GitHub credentials. - Create your own `codecov.io` to help with testing. - The PR you create will show a report for the code coverage for the file you have added. You can also use that as a guide.
non_process
test src lib resolvers subscription directmessagechat ts please coordinate issue assignment and pr reviews with the contributors listed in this issue the talawa api code base needs to be reliable this means we need to have test code coverage tests need to be written for file src lib resolvers subscription directmessagechat ts we will need the api to be refactored for all methods classes and or functions found in this file for testing to be correctly executed when complete all all methods classes and or functions in the refactored file will need to be tested these tests must be placed in a single file with the name tests resolvers subscription directmessagechat spec ts you may need to create the appropriate directory structure to do this important please refer to the parent issue on how to implement these tests correctly pr acceptance criteria when complete this file must show coverage when merged into the code base this will be clearly visible when you submit your pr login using your github credentials create your own codecov io to help with testing the pr you create will show a report for the code coverage for the file you have added you can also use that as a guide
0
16,285
20,907,965,534
IssuesEvent
2022-03-24 05:48:19
dlehdgus9887/paper-log-own
https://api.github.com/repos/dlehdgus9887/paper-log-own
opened
Deep Gaussian Process
2017 Gaussian Process
## 어떤 내용의 논문인가요? 👋 하나의 gasussian process에서 그치는 것이 아니라 다양한 gaussian process를 쌓아서 서로의 출력과 입력을 활용한다. ## Abstract (요약) 🕵🏻‍♂️ Deep Gaussian ProcessesAndreas C. DamianouNeil D. LawrenceDept. of Computer Science & Sheffield Institute for Translational Neuroscience,University of Sheffield, UKAbstractIn this paper we introduce deep Gaussian process(GP) models. Deep GPs are a deep belief net-work based on Gaussian process mappings. Thedata is modeled as the output of a multivariateGP. The inputs to that Gaussian process are thengoverned by another GP. A single layer model isequivalent to a standard GP or the GP latent vari-able model (GP-LVM). We perform inference inthe model by approximate variational marginal-ization. This results in a strict lower bound on themarginal likelihood of the model which we usefor model selection (number of layers and nodesper layer). Deep belief networks are typically ap-plied to relatively large data sets using stochas-tic gradient descent for optimization. Our fullyBayesian treatment allows for the application ofdeep models even when data is scarce. Model se-lection by our variational bound shows that a fivelayer hierarchy is justified even when modellinga digit data set containing only 150 examples ## 이 논문을 읽어서 무엇을 배울 수 있는지 알려주세요! 🤔 Gaussian process란 함수의 분포를 추정하는 것이고 이것을 수직적으로 쌓아나가는 것이 논리적이다. ## 같이 읽어보면 좋을 만한 글이나 이슈가 있을까요? 만약에 있다면 자유롭게 작성해 주세요! ## 레퍼런스의 URL을 알려주세요! 🔗 http://proceedings.mlr.press/v31/damianou13a
1.0
Deep Gaussian Process - ## 어떤 내용의 논문인가요? 👋 하나의 gasussian process에서 그치는 것이 아니라 다양한 gaussian process를 쌓아서 서로의 출력과 입력을 활용한다. ## Abstract (요약) 🕵🏻‍♂️ Deep Gaussian ProcessesAndreas C. DamianouNeil D. LawrenceDept. of Computer Science & Sheffield Institute for Translational Neuroscience,University of Sheffield, UKAbstractIn this paper we introduce deep Gaussian process(GP) models. Deep GPs are a deep belief net-work based on Gaussian process mappings. Thedata is modeled as the output of a multivariateGP. The inputs to that Gaussian process are thengoverned by another GP. A single layer model isequivalent to a standard GP or the GP latent vari-able model (GP-LVM). We perform inference inthe model by approximate variational marginal-ization. This results in a strict lower bound on themarginal likelihood of the model which we usefor model selection (number of layers and nodesper layer). Deep belief networks are typically ap-plied to relatively large data sets using stochas-tic gradient descent for optimization. Our fullyBayesian treatment allows for the application ofdeep models even when data is scarce. Model se-lection by our variational bound shows that a fivelayer hierarchy is justified even when modellinga digit data set containing only 150 examples ## 이 논문을 읽어서 무엇을 배울 수 있는지 알려주세요! 🤔 Gaussian process란 함수의 분포를 추정하는 것이고 이것을 수직적으로 쌓아나가는 것이 논리적이다. ## 같이 읽어보면 좋을 만한 글이나 이슈가 있을까요? 만약에 있다면 자유롭게 작성해 주세요! ## 레퍼런스의 URL을 알려주세요! 🔗 http://proceedings.mlr.press/v31/damianou13a
process
deep gaussian process 어떤 내용의 논문인가요 👋 하나의 gasussian process에서 그치는 것이 아니라 다양한 gaussian process를 쌓아서 서로의 출력과 입력을 활용한다 abstract 요약 🕵🏻‍♂️ deep gaussian processesandreas c damianouneil d lawrencedept of computer science sheffield institute for translational neuroscience university of sheffield ukabstractin this paper we introduce deep gaussian process gp models deep gps are a deep belief net work based on gaussian process mappings thedata is modeled as the output of a multivariategp the inputs to that gaussian process are thengoverned by another gp a single layer model isequivalent to a standard gp or the gp latent vari able model gp lvm we perform inference inthe model by approximate variational marginal ization this results in a strict lower bound on themarginal likelihood of the model which we usefor model selection number of layers and nodesper layer deep belief networks are typically ap plied to relatively large data sets using stochas tic gradient descent for optimization our fullybayesian treatment allows for the application ofdeep models even when data is scarce model se lection by our variational bound shows that a fivelayer hierarchy is justified even when modellinga digit data set containing only examples 이 논문을 읽어서 무엇을 배울 수 있는지 알려주세요 🤔 gaussian process란 함수의 분포를 추정하는 것이고 이것을 수직적으로 쌓아나가는 것이 논리적이다 같이 읽어보면 좋을 만한 글이나 이슈가 있을까요 만약에 있다면 자유롭게 작성해 주세요 레퍼런스의 url을 알려주세요 🔗
1
13,641
16,329,736,369
IssuesEvent
2021-05-12 07:43:14
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
`connect_timeout` does not work correctly for Introspection
bug/1-repro-available kind/bug process/candidate team/migrations topic: connection-url topic: introspection
### Bug description It seems that for Introspection Engine the `connect_timeout` connection URL parameter does not work properly. To test this, I have a locally running PostgreSQL server at port 5432. I am using a connection URL that does not (!) include the correct credentials, so when connecting to the database server you should get the following error: ``` Error: P1000: Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. ``` To be able to trigger the timeout when creating a connection (which usually only takes a few ms), I am using https://www.netlimiter.com to throttle the network for the engine binaries to 1B/s which makes the calls _really_ slow. I am using Windows Powershell's `Measure-Command { ... } | Select-Object -Property TotalSeconds` to measure command duration and output the duration in seconds. (As I am also showing that for the Query Engine the `connect_timeout` works fine, I also added `pool_timeout=0` to the connection URL to make sure there is no pool timeout. This is irrelevant for ME and IE of course.) Below I run `npx prisma introspect`, but also `npx prisma db push` and a `node script.js` for each configuration to show that the Migration and Query Engine are not affected - only Introspection Engine is. <details> <summary>Content of `script.js`</summary> ``` const { PrismaClient } = require('@prisma/client') const prisma = new PrismaClient() // A `main` function so that we can use async/await async function main() { // Seed the database with users and posts const user1 = await prisma.user.create({ data: { email: 'alice@prisma.io', name: 'Alice', }, }) } main() .catch((e) => { console.error(e) process.exit(1) }) .finally(async () => { await prisma.$disconnect() }) ``` </details> ## Reproduction ### Control, with no throttling and no custom `connect_timeout` (default = 5) ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1000 Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. TotalSeconds ------------ 1.4380423 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1000: Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. TotalSeconds ------------ 1.5172195 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 0.4597608 ``` ### Throttling + No custom `connect_timeout` (default = 5) ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 6.2946373 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 6.4426992 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 5.3147573 ``` ### Throttling + `connect_timeout=1` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 2.2411517 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 2.4388567 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 1.2970396 ``` ### Throttling + `connect_timeout=5` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. ``` ``` TotalSeconds ------------ 6.2502259 PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 6.4696914 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 5.2912967 ``` ### Throttling + `connect_timeout=10` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.2784433 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.3936712 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 10.2934561 ``` ### Throttling + `connect_timeout=15` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.290506 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 16.8142782 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 15.2977498 ``` ### Throttling + `connect_timeout=30` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.367393 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 32.3986191 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 30.2955554 ``` ### Throttling + `connect_timeout=60` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 12.0734391 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 62.1473232 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 60.6906105 ``` So it seems that up to values of `connect_timeout=10` Introspection Engine behaves as expected, but then it does not any more.
1.0
`connect_timeout` does not work correctly for Introspection - ### Bug description It seems that for Introspection Engine the `connect_timeout` connection URL parameter does not work properly. To test this, I have a locally running PostgreSQL server at port 5432. I am using a connection URL that does not (!) include the correct credentials, so when connecting to the database server you should get the following error: ``` Error: P1000: Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. ``` To be able to trigger the timeout when creating a connection (which usually only takes a few ms), I am using https://www.netlimiter.com to throttle the network for the engine binaries to 1B/s which makes the calls _really_ slow. I am using Windows Powershell's `Measure-Command { ... } | Select-Object -Property TotalSeconds` to measure command duration and output the duration in seconds. (As I am also showing that for the Query Engine the `connect_timeout` works fine, I also added `pool_timeout=0` to the connection URL to make sure there is no pool timeout. This is irrelevant for ME and IE of course.) Below I run `npx prisma introspect`, but also `npx prisma db push` and a `node script.js` for each configuration to show that the Migration and Query Engine are not affected - only Introspection Engine is. <details> <summary>Content of `script.js`</summary> ``` const { PrismaClient } = require('@prisma/client') const prisma = new PrismaClient() // A `main` function so that we can use async/await async function main() { // Seed the database with users and posts const user1 = await prisma.user.create({ data: { email: 'alice@prisma.io', name: 'Alice', }, }) } main() .catch((e) => { console.error(e) process.exit(1) }) .finally(async () => { await prisma.$disconnect() }) ``` </details> ## Reproduction ### Control, with no throttling and no custom `connect_timeout` (default = 5) ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1000 Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. TotalSeconds ------------ 1.4380423 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1000: Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. TotalSeconds ------------ 1.5172195 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Authentication failed against database server at `localhost`, the provided database credentials for `johndoe` are not valid. Please make sure to provide valid database credentials for the database server at `localhost`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 0.4597608 ``` ### Throttling + No custom `connect_timeout` (default = 5) ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 6.2946373 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 6.4426992 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 5.3147573 ``` ### Throttling + `connect_timeout=1` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 2.2411517 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 2.4388567 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 1.2970396 ``` ### Throttling + `connect_timeout=5` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. ``` ``` TotalSeconds ------------ 6.2502259 PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 6.4696914 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 5.2912967 ``` ### Throttling + `connect_timeout=10` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.2784433 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.3936712 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 10.2934561 ``` ### Throttling + `connect_timeout=15` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.290506 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 16.8142782 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 15.2977498 ``` ### Throttling + `connect_timeout=30` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 11.367393 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 32.3986191 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 30.2955554 ``` ### Throttling + `connect_timeout=60` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma introspect } | Select-Object -Property TotalSeconds Error: P1001 Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 12.0734391 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { npx prisma db push } | Select-Object -Property TotalSeconds Error: P1001: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. TotalSeconds ------------ 62.1473232 ``` ``` PS C:\Users\Jan\Documents\throwaway\timeoutTest> Measure-Command { node script.js } | Select-Object -Property TotalSeconds PrismaClientInitializationError2 [PrismaClientInitializationError]: Invalid `prisma.user.create()` invocation: Can't reach database server at `localhost`:`5432` Please make sure your database server is running at `localhost`:`5432`. at cb (C:\Users\Jan\Documents\throwaway\timeoutTest\node_modules\@prisma\client\runtime\index.js:35104:17) at processTicksAndRejections (internal/process/task_queues.js:93:5) at async main (C:\Users\Jan\Documents\throwaway\timeoutTest\script.js:8:17) { clientVersion: '2.22.1', errorCode: undefined } TotalSeconds ------------ 60.6906105 ``` So it seems that up to values of `connect_timeout=10` Introspection Engine behaves as expected, but then it does not any more.
process
connect timeout does not work correctly for introspection bug description it seems that for introspection engine the connect timeout connection url parameter does not work properly to test this i have a locally running postgresql server at port i am using a connection url that does not include the correct credentials so when connecting to the database server you should get the following error error authentication failed against database server at localhost the provided database credentials for johndoe are not valid please make sure to provide valid database credentials for the database server at localhost to be able to trigger the timeout when creating a connection which usually only takes a few ms i am using to throttle the network for the engine binaries to s which makes the calls really slow i am using windows powershell s measure command select object property totalseconds to measure command duration and output the duration in seconds as i am also showing that for the query engine the connect timeout works fine i also added pool timeout to the connection url to make sure there is no pool timeout this is irrelevant for me and ie of course below i run npx prisma introspect but also npx prisma db push and a node script js for each configuration to show that the migration and query engine are not affected only introspection engine is content of script js const prismaclient require prisma client const prisma new prismaclient a main function so that we can use async await async function main seed the database with users and posts const await prisma user create data email alice prisma io name alice main catch e console error e process exit finally async await prisma disconnect reproduction control with no throttling and no custom connect timeout default ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error authentication failed against database server at localhost the provided database credentials for johndoe are not valid please make sure to provide valid database credentials for the database server at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error authentication failed against database server at localhost the provided database credentials for johndoe are not valid please make sure to provide valid database credentials for the database server at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation authentication failed against database server at localhost the provided database credentials for johndoe are not valid please make sure to provide valid database credentials for the database server at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds throttling no custom connect timeout default ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation can t reach database server at localhost please make sure your database server is running at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds throttling connect timeout ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation can t reach database server at localhost please make sure your database server is running at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds throttling connect timeout ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation can t reach database server at localhost please make sure your database server is running at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds throttling connect timeout ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation can t reach database server at localhost please make sure your database server is running at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds throttling connect timeout ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation can t reach database server at localhost please make sure your database server is running at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds throttling connect timeout ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation can t reach database server at localhost please make sure your database server is running at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds throttling connect timeout ps c users jan documents throwaway timeouttest measure command npx prisma introspect select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command npx prisma db push select object property totalseconds error can t reach database server at localhost please make sure your database server is running at localhost totalseconds ps c users jan documents throwaway timeouttest measure command node script js select object property totalseconds invalid prisma user create invocation can t reach database server at localhost please make sure your database server is running at localhost at cb c users jan documents throwaway timeouttest node modules prisma client runtime index js at processticksandrejections internal process task queues js at async main c users jan documents throwaway timeouttest script js clientversion errorcode undefined totalseconds so it seems that up to values of connect timeout introspection engine behaves as expected but then it does not any more
1
363,159
10,738,514,807
IssuesEvent
2019-10-29 14:54:21
alerj/procuradoria
https://api.github.com/repos/alerj/procuradoria
closed
Refactor de pareceres
priority 1
- Todos podem entrar no sistema de pareceres - Todos podem criar/editar pareceres, exceto estagiários - Uma vez que foi colocado um arquivo (pdf ou doc), ele não pode mais ser alterado - Tirar a exibição do form de Opinions dinamicamente - Criar campos para os pareceres: - [x] is_active (boolean) - [x] approve_option_id (Criar uma tabela auxiliar approve_options - SIM / NÃO / NÃO HÁ INFORMAÇÃO)
1.0
Refactor de pareceres - - Todos podem entrar no sistema de pareceres - Todos podem criar/editar pareceres, exceto estagiários - Uma vez que foi colocado um arquivo (pdf ou doc), ele não pode mais ser alterado - Tirar a exibição do form de Opinions dinamicamente - Criar campos para os pareceres: - [x] is_active (boolean) - [x] approve_option_id (Criar uma tabela auxiliar approve_options - SIM / NÃO / NÃO HÁ INFORMAÇÃO)
non_process
refactor de pareceres todos podem entrar no sistema de pareceres todos podem criar editar pareceres exceto estagiários uma vez que foi colocado um arquivo pdf ou doc ele não pode mais ser alterado tirar a exibição do form de opinions dinamicamente criar campos para os pareceres is active boolean approve option id criar uma tabela auxiliar approve options sim não não há informação
0
457,697
13,160,322,959
IssuesEvent
2020-08-10 17:22:49
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
`cattle-cluster-agent` in downstream cluster keeps crashing because of segmentation fault
[zube]: Next Up alpha-priority/0 kind/bug-qa status/blocker
<!-- Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase. --> **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - install `rancher:master-3420-head` in HA mode - go to dashboard -> app -> enable the ECM feature - go to ember UI - import an RKE cluster ( in my case, 1 node, v1.18.3) **Result:** - the cluster fails at health checking <img width="1445" alt="Screen Shot 2020-08-06 at 11 22 03 AM" src="https://user-images.githubusercontent.com/6218999/89568354-0a10c380-d7d8-11ea-9fe8-8c82b585e9c5.png"> - in the cluster, the pod `cattle-cluster-agent` is in a crashing loop ``` NAMESPACE NAME READY STATUS RESTARTS AGE cattle-system cattle-cluster-agent-bcb94b77c-dtx2t 0/1 CrashLoopBackOff 5 5m14s ``` - here are the logs of the pod ``` INFO: Environment: CATTLE_ADDRESS=10.42.0.6 CATTLE_CA_CHECKSUM=295b5528d021581b9403663361acc7bfd7ac5c8669c30ca7be307a69e7d96789 CATTLE_CLUSTER=true CATTLE_FEATURES=,multi-cluster-management=disable CATTLE_INTERNAL_ADDRESS= CATTLE_IS_RKE=false CATTLE_K8S_MANAGED=true CATTLE_NODE_NAME=cattle-cluster-agent-bcb94b77c-dtx2t CATTLE_SERVER=https://jiaqi02391.whatever.domain.com INFO: Using resolv.conf: nameserver 10.43.0.10 search cattle-system.svc.cluster.local svc.cluster.local cluster.local us-west-1.compute.internal options ndots:5 INFO: https://jiaqi02391.whatever.domain.com/ping is accessible INFO: jiaqi02391.whatever.domain.com resolves to 3.129.28.202 INFO: Value from https://jiaqi02391.whatever.domain.com/v3/settings/cacerts is an x509 certificate time="2020-08-06T18:19:20Z" level=error msg="failed to apply feature args: feature argument should be of the form \"features=feature1=bool,feature2=bool\"" time="2020-08-06T18:19:20Z" level=info msg="Rancher agent version 9b0dd20b7 is starting" time="2020-08-06T18:19:20Z" level=info msg="Listening on /tmp/log.sock" time="2020-08-06T18:19:20Z" level=info msg="Connecting to wss://jiaqi02391.whatever.domain.com/v3/connect/register with token 4qhngkt5pb6xz7j44ml5phdnqwrpkfcqks5wlx6vf7z4t42gdxqzt5" time="2020-08-06T18:19:20Z" level=info msg="Connecting to proxy" url="wss://jiaqi02391.whatever.domain.com/v3/connect/register" time="2020-08-06T18:19:20Z" level=info msg="Starting steve" time="2020-08-06T18:19:20Z" level=info msg="Updating CRD releases.catalog.cattle.io" I0806 18:19:21.572087 1 leaderelection.go:242] attempting to acquire leader lease kube-system/cattle-controllers... time="2020-08-06T18:19:21Z" level=info msg="Starting management.cattle.io/v3, Kind=Setting controller" time="2020-08-06T18:19:21Z" level=info msg="Starting apiregistration.k8s.io/v1, Kind=APIService controller" time="2020-08-06T18:19:21Z" level=info msg="Starting management.cattle.io/v3, Kind=Cluster controller" time="2020-08-06T18:19:21Z" level=info msg="Starting /v1, Kind=ConfigMap controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=Role controller" time="2020-08-06T18:19:21Z" level=info msg="Starting catalog.cattle.io/v1, Kind=Repo controller" time="2020-08-06T18:19:21Z" level=info msg="Starting catalog.cattle.io/v1, Kind=ClusterRepo controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=RoleBinding controller" time="2020-08-06T18:19:21Z" level=info msg="Starting management.cattle.io/v3, Kind=Feature controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRole controller" time="2020-08-06T18:19:21Z" level=info msg="Starting apiextensions.k8s.io/v1beta1, Kind=CustomResourceDefinition controller" time="2020-08-06T18:19:21Z" level=info msg="Starting /v1, Kind=Secret controller" I0806 18:19:21.610268 1 leaderelection.go:252] successfully acquired lease kube-system/cattle-controllers panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x290 pc=0x1d9d6c1] goroutine 196 [running]: github.com/rancher/rancher/pkg/auth.(*Server).OnLeader(0xc000756700, 0x4ca3140, 0xc000010c80, 0x1, 0x1) /go/src/github.com/rancher/rancher/pkg/auth/server.go:118 +0x71 github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader.(*Manager).OnLeader.func1(0xc001157c80, 0xc000332480) /go/src/github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader/manager.go:51 +0xcc created by github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader.(*Manager).OnLeader /go/src/github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader/manager.go:48 +0x49 ``` **Other details that may be helpful:** The same error happens to custom clusters and rancher-provisioned clusters, too. Update: - see the same error in rancher single-install setup
1.0
`cattle-cluster-agent` in downstream cluster keeps crashing because of segmentation fault - <!-- Please search for existing issues first, then read https://rancher.com/docs/rancher/v2.x/en/contributing/#bugs-issues-or-questions to see what we expect in an issue For security issues, please email security@rancher.com instead of posting a public issue in GitHub. You may (but are not required to) use the GPG key located on Keybase. --> **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** - install `rancher:master-3420-head` in HA mode - go to dashboard -> app -> enable the ECM feature - go to ember UI - import an RKE cluster ( in my case, 1 node, v1.18.3) **Result:** - the cluster fails at health checking <img width="1445" alt="Screen Shot 2020-08-06 at 11 22 03 AM" src="https://user-images.githubusercontent.com/6218999/89568354-0a10c380-d7d8-11ea-9fe8-8c82b585e9c5.png"> - in the cluster, the pod `cattle-cluster-agent` is in a crashing loop ``` NAMESPACE NAME READY STATUS RESTARTS AGE cattle-system cattle-cluster-agent-bcb94b77c-dtx2t 0/1 CrashLoopBackOff 5 5m14s ``` - here are the logs of the pod ``` INFO: Environment: CATTLE_ADDRESS=10.42.0.6 CATTLE_CA_CHECKSUM=295b5528d021581b9403663361acc7bfd7ac5c8669c30ca7be307a69e7d96789 CATTLE_CLUSTER=true CATTLE_FEATURES=,multi-cluster-management=disable CATTLE_INTERNAL_ADDRESS= CATTLE_IS_RKE=false CATTLE_K8S_MANAGED=true CATTLE_NODE_NAME=cattle-cluster-agent-bcb94b77c-dtx2t CATTLE_SERVER=https://jiaqi02391.whatever.domain.com INFO: Using resolv.conf: nameserver 10.43.0.10 search cattle-system.svc.cluster.local svc.cluster.local cluster.local us-west-1.compute.internal options ndots:5 INFO: https://jiaqi02391.whatever.domain.com/ping is accessible INFO: jiaqi02391.whatever.domain.com resolves to 3.129.28.202 INFO: Value from https://jiaqi02391.whatever.domain.com/v3/settings/cacerts is an x509 certificate time="2020-08-06T18:19:20Z" level=error msg="failed to apply feature args: feature argument should be of the form \"features=feature1=bool,feature2=bool\"" time="2020-08-06T18:19:20Z" level=info msg="Rancher agent version 9b0dd20b7 is starting" time="2020-08-06T18:19:20Z" level=info msg="Listening on /tmp/log.sock" time="2020-08-06T18:19:20Z" level=info msg="Connecting to wss://jiaqi02391.whatever.domain.com/v3/connect/register with token 4qhngkt5pb6xz7j44ml5phdnqwrpkfcqks5wlx6vf7z4t42gdxqzt5" time="2020-08-06T18:19:20Z" level=info msg="Connecting to proxy" url="wss://jiaqi02391.whatever.domain.com/v3/connect/register" time="2020-08-06T18:19:20Z" level=info msg="Starting steve" time="2020-08-06T18:19:20Z" level=info msg="Updating CRD releases.catalog.cattle.io" I0806 18:19:21.572087 1 leaderelection.go:242] attempting to acquire leader lease kube-system/cattle-controllers... time="2020-08-06T18:19:21Z" level=info msg="Starting management.cattle.io/v3, Kind=Setting controller" time="2020-08-06T18:19:21Z" level=info msg="Starting apiregistration.k8s.io/v1, Kind=APIService controller" time="2020-08-06T18:19:21Z" level=info msg="Starting management.cattle.io/v3, Kind=Cluster controller" time="2020-08-06T18:19:21Z" level=info msg="Starting /v1, Kind=ConfigMap controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=Role controller" time="2020-08-06T18:19:21Z" level=info msg="Starting catalog.cattle.io/v1, Kind=Repo controller" time="2020-08-06T18:19:21Z" level=info msg="Starting catalog.cattle.io/v1, Kind=ClusterRepo controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=RoleBinding controller" time="2020-08-06T18:19:21Z" level=info msg="Starting management.cattle.io/v3, Kind=Feature controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRoleBinding controller" time="2020-08-06T18:19:21Z" level=info msg="Starting rbac.authorization.k8s.io/v1, Kind=ClusterRole controller" time="2020-08-06T18:19:21Z" level=info msg="Starting apiextensions.k8s.io/v1beta1, Kind=CustomResourceDefinition controller" time="2020-08-06T18:19:21Z" level=info msg="Starting /v1, Kind=Secret controller" I0806 18:19:21.610268 1 leaderelection.go:252] successfully acquired lease kube-system/cattle-controllers panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x1 addr=0x290 pc=0x1d9d6c1] goroutine 196 [running]: github.com/rancher/rancher/pkg/auth.(*Server).OnLeader(0xc000756700, 0x4ca3140, 0xc000010c80, 0x1, 0x1) /go/src/github.com/rancher/rancher/pkg/auth/server.go:118 +0x71 github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader.(*Manager).OnLeader.func1(0xc001157c80, 0xc000332480) /go/src/github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader/manager.go:51 +0xcc created by github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader.(*Manager).OnLeader /go/src/github.com/rancher/rancher/vendor/github.com/rancher/wrangler/pkg/leader/manager.go:48 +0x49 ``` **Other details that may be helpful:** The same error happens to custom clusters and rancher-provisioned clusters, too. Update: - see the same error in rancher single-install setup
non_process
cattle cluster agent in downstream cluster keeps crashing because of segmentation fault please search for existing issues first then read to see what we expect in an issue for security issues please email security rancher com instead of posting a public issue in github you may but are not required to use the gpg key located on keybase what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible install rancher master head in ha mode go to dashboard app enable the ecm feature go to ember ui import an rke cluster in my case node result the cluster fails at health checking img width alt screen shot at am src in the cluster the pod cattle cluster agent is in a crashing loop namespace name ready status restarts age cattle system cattle cluster agent crashloopbackoff here are the logs of the pod info environment cattle address cattle ca checksum cattle cluster true cattle features multi cluster management disable cattle internal address cattle is rke false cattle managed true cattle node name cattle cluster agent cattle server info using resolv conf nameserver search cattle system svc cluster local svc cluster local cluster local us west compute internal options ndots info is accessible info whatever domain com resolves to info value from is an certificate time level error msg failed to apply feature args feature argument should be of the form features bool bool time level info msg rancher agent version is starting time level info msg listening on tmp log sock time level info msg connecting to wss whatever domain com connect register with token time level info msg connecting to proxy url wss whatever domain com connect register time level info msg starting steve time level info msg updating crd releases catalog cattle io leaderelection go attempting to acquire leader lease kube system cattle controllers time level info msg starting management cattle io kind setting controller time level info msg starting apiregistration io kind apiservice controller time level info msg starting management cattle io kind cluster controller time level info msg starting kind configmap controller time level info msg starting rbac authorization io kind role controller time level info msg starting catalog cattle io kind repo controller time level info msg starting catalog cattle io kind clusterrepo controller time level info msg starting rbac authorization io kind rolebinding controller time level info msg starting management cattle io kind feature controller time level info msg starting rbac authorization io kind clusterrolebinding controller time level info msg starting rbac authorization io kind clusterrole controller time level info msg starting apiextensions io kind customresourcedefinition controller time level info msg starting kind secret controller leaderelection go successfully acquired lease kube system cattle controllers panic runtime error invalid memory address or nil pointer dereference goroutine github com rancher rancher pkg auth server onleader go src github com rancher rancher pkg auth server go github com rancher rancher vendor github com rancher wrangler pkg leader manager onleader go src github com rancher rancher vendor github com rancher wrangler pkg leader manager go created by github com rancher rancher vendor github com rancher wrangler pkg leader manager onleader go src github com rancher rancher vendor github com rancher wrangler pkg leader manager go other details that may be helpful the same error happens to custom clusters and rancher provisioned clusters too update see the same error in rancher single install setup
0
112,555
24,293,123,962
IssuesEvent
2022-09-29 07:57:18
Onelinerhub/onelinerhub
https://api.github.com/repos/Onelinerhub/onelinerhub
closed
Short solution needed: "php gd unicode" (php-gd)
help wanted good first issue code php-gd
Please help us write most modern and shortest code solution for this issue: **php gd unicode** (technology: [php-gd](https://onelinerhub.com/php-gd)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution. 3. Link to this issue in comments of pull request.
1.0
Short solution needed: "php gd unicode" (php-gd) - Please help us write most modern and shortest code solution for this issue: **php gd unicode** (technology: [php-gd](https://onelinerhub.com/php-gd)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution. 3. Link to this issue in comments of pull request.
non_process
short solution needed php gd unicode php gd please help us write most modern and shortest code solution for this issue php gd unicode technology fast way just write the code solution in the comments prefered way create with a new code file inside don t forget to explain solution link to this issue in comments of pull request
0
26,095
26,407,411,491
IssuesEvent
2023-01-13 09:17:31
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
F5 shortcut gets eaten when numeral fields are selected in the inspector
enhancement topic:editor confirmed usability
**Godot version:** 3.3.1.stable **OS/device including version:** Windows 10 Pro Version 20H2 (OS Build 19042.985) **Issue description:** I wanted to test my game after making some changes in the inspector. I pressed F5 and nothing happened. I expected F5 to launch the game. After some experimentation I found this bug occurs only when I have a numeral field selected in the inspector. **Steps to reproduce:** Place a Node2D (for example) in the scene. Click on the Transform Position.x field in the inspector and type in a value. DO NOT PRESS ENTER OR CLICK AWAY. Either leave the field selected or press tab to select the next field (Position.y). Press F5. Nothing will happen. Click out of the field or press Enter to deselect it and press F5. The game will launch. This only seems to happen when editing numeric fields. F5 functions as expected when selecting checkboxes or editing text/string fields. **Minimal reproduction project:** [BugReport.zip](https://github.com/godotengine/godot/files/6506129/BugReport.zip)
True
F5 shortcut gets eaten when numeral fields are selected in the inspector - **Godot version:** 3.3.1.stable **OS/device including version:** Windows 10 Pro Version 20H2 (OS Build 19042.985) **Issue description:** I wanted to test my game after making some changes in the inspector. I pressed F5 and nothing happened. I expected F5 to launch the game. After some experimentation I found this bug occurs only when I have a numeral field selected in the inspector. **Steps to reproduce:** Place a Node2D (for example) in the scene. Click on the Transform Position.x field in the inspector and type in a value. DO NOT PRESS ENTER OR CLICK AWAY. Either leave the field selected or press tab to select the next field (Position.y). Press F5. Nothing will happen. Click out of the field or press Enter to deselect it and press F5. The game will launch. This only seems to happen when editing numeric fields. F5 functions as expected when selecting checkboxes or editing text/string fields. **Minimal reproduction project:** [BugReport.zip](https://github.com/godotengine/godot/files/6506129/BugReport.zip)
non_process
shortcut gets eaten when numeral fields are selected in the inspector godot version stable os device including version windows pro version os build issue description i wanted to test my game after making some changes in the inspector i pressed and nothing happened i expected to launch the game after some experimentation i found this bug occurs only when i have a numeral field selected in the inspector steps to reproduce place a for example in the scene click on the transform position x field in the inspector and type in a value do not press enter or click away either leave the field selected or press tab to select the next field position y press nothing will happen click out of the field or press enter to deselect it and press the game will launch this only seems to happen when editing numeric fields functions as expected when selecting checkboxes or editing text string fields minimal reproduction project
0
12,088
14,740,060,691
IssuesEvent
2021-01-07 08:26:51
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Memphis - SA Billing - Late Fee Account List
anc-process anp-important ant-bug has attachment
In GitLab by @kdjstudios on Oct 3, 2018, 11:04 [Memphis.xlsx](/uploads/aa62ea4765bfff350f96c711908dba08/Memphis.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-31732/conversation
1.0
Memphis - SA Billing - Late Fee Account List - In GitLab by @kdjstudios on Oct 3, 2018, 11:04 [Memphis.xlsx](/uploads/aa62ea4765bfff350f96c711908dba08/Memphis.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-31732/conversation
process
memphis sa billing late fee account list in gitlab by kdjstudios on oct uploads memphis xlsx hd
1
22,671
31,896,143,367
IssuesEvent
2023-09-18 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Mon, 18 Sep 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Deformable Neural Radiance Fields using RGB and Event Cameras - **Authors:** Qi Ma, Danda Pani Paudel, Ajad Chhatkuli, Luc Van Gool - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08416 - **Pdf link:** https://arxiv.org/pdf/2309.08416 - **Abstract** Modeling Neural Radiance Fields for fast-moving deformable objects from visual data alone is a challenging problem. A major issue arises due to the high deformation and low acquisition rates. To address this problem, we propose to use event cameras that offer very fast acquisition of visual change in an asynchronous manner. In this work, we develop a novel method to model the deformable neural radiance fields using RGB and event cameras. The proposed method uses the asynchronous stream of events and calibrated sparse RGB frames. In our setup, the camera pose at the individual events required to integrate them into the radiance fields remains unknown. Our method jointly optimizes these poses and the radiance field. This happens efficiently by leveraging the collection of events at once and actively sampling the events during learning. Experiments conducted on both realistically rendered graphics and real-world datasets demonstrate a significant benefit of the proposed method over the state-of-the-art and the compared baseline. This shows a promising direction for modeling deformable neural radiance fields in real-world dynamic scenes. ### Robust e-NeRF: NeRF from Sparse & Noisy Events under Non-Uniform Motion - **Authors:** Weng Fei Low, Gim Hee Lee - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2309.08596 - **Pdf link:** https://arxiv.org/pdf/2309.08596 - **Abstract** Event cameras offer many advantages over standard cameras due to their distinctive principle of operation: low power, low latency, high temporal resolution and high dynamic range. Nonetheless, the success of many downstream visual applications also hinges on an efficient and effective scene representation, where Neural Radiance Field (NeRF) is seen as the leading candidate. Such promise and potential of event cameras and NeRF inspired recent works to investigate on the reconstruction of NeRF from moving event cameras. However, these works are mainly limited in terms of the dependence on dense and low-noise event streams, as well as generalization to arbitrary contrast threshold values and camera speed profiles. In this work, we propose Robust e-NeRF, a novel method to directly and robustly reconstruct NeRFs from moving event cameras under various real-world conditions, especially from sparse and noisy events generated under non-uniform motion. It consists of two key components: a realistic event generation model that accounts for various intrinsic parameters (e.g. time-independent, asymmetric threshold and refractory period) and non-idealities (e.g. pixel-to-pixel threshold variation), as well as a complementary pair of normalized reconstruction losses that can effectively generalize to arbitrary speed profiles and intrinsic parameter values without such prior knowledge. Experiments on real and novel realistically simulated sequences verify our effectiveness. Our code, synthetic dataset and improved event simulator are public. ## Keyword: event camera ### Deformable Neural Radiance Fields using RGB and Event Cameras - **Authors:** Qi Ma, Danda Pani Paudel, Ajad Chhatkuli, Luc Van Gool - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08416 - **Pdf link:** https://arxiv.org/pdf/2309.08416 - **Abstract** Modeling Neural Radiance Fields for fast-moving deformable objects from visual data alone is a challenging problem. A major issue arises due to the high deformation and low acquisition rates. To address this problem, we propose to use event cameras that offer very fast acquisition of visual change in an asynchronous manner. In this work, we develop a novel method to model the deformable neural radiance fields using RGB and event cameras. The proposed method uses the asynchronous stream of events and calibrated sparse RGB frames. In our setup, the camera pose at the individual events required to integrate them into the radiance fields remains unknown. Our method jointly optimizes these poses and the radiance field. This happens efficiently by leveraging the collection of events at once and actively sampling the events during learning. Experiments conducted on both realistically rendered graphics and real-world datasets demonstrate a significant benefit of the proposed method over the state-of-the-art and the compared baseline. This shows a promising direction for modeling deformable neural radiance fields in real-world dynamic scenes. ### Robust e-NeRF: NeRF from Sparse & Noisy Events under Non-Uniform Motion - **Authors:** Weng Fei Low, Gim Hee Lee - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2309.08596 - **Pdf link:** https://arxiv.org/pdf/2309.08596 - **Abstract** Event cameras offer many advantages over standard cameras due to their distinctive principle of operation: low power, low latency, high temporal resolution and high dynamic range. Nonetheless, the success of many downstream visual applications also hinges on an efficient and effective scene representation, where Neural Radiance Field (NeRF) is seen as the leading candidate. Such promise and potential of event cameras and NeRF inspired recent works to investigate on the reconstruction of NeRF from moving event cameras. However, these works are mainly limited in terms of the dependence on dense and low-noise event streams, as well as generalization to arbitrary contrast threshold values and camera speed profiles. In this work, we propose Robust e-NeRF, a novel method to directly and robustly reconstruct NeRFs from moving event cameras under various real-world conditions, especially from sparse and noisy events generated under non-uniform motion. It consists of two key components: a realistic event generation model that accounts for various intrinsic parameters (e.g. time-independent, asymmetric threshold and refractory period) and non-idealities (e.g. pixel-to-pixel threshold variation), as well as a complementary pair of normalized reconstruction losses that can effectively generalize to arbitrary speed profiles and intrinsic parameter values without such prior knowledge. Experiments on real and novel realistically simulated sequences verify our effectiveness. Our code, synthetic dataset and improved event simulator are public. ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Depth Estimation from a Single Optical Encoded Image using a Learned Colored-Coded Aperture - **Authors:** Jhon Lopez, Edwin Vargas, Henry Arguello - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08033 - **Pdf link:** https://arxiv.org/pdf/2309.08033 - **Abstract** Depth estimation from a single image of a conventional camera is a challenging task since depth cues are lost during the acquisition process. State-of-the-art approaches improve the discrimination between different depths by introducing a binary-coded aperture (CA) in the lens aperture that generates different coded blur patterns at different depths. Color-coded apertures (CCA) can also produce color misalignment in the captured image which can be utilized to estimate disparity. Leveraging advances in deep learning, more recent works have explored the data-driven design of a diffractive optical element (DOE) for encoding depth information through chromatic aberrations. However, compared with binary CA or CCA, DOEs are more expensive to fabricate and require high-precision devices. Different from previous CCA-based approaches that employ few basic colors, in this work we propose a CCA with a greater number of color filters and richer spectral information to optically encode relevant depth information in a single snapshot. Furthermore, we propose to jointly learn the color-coded aperture (CCA) pattern and a convolutional neural network (CNN) to retrieve depth information by using an end-to-end optimization approach. We demonstrate through different experiments on three different data sets that the designed color-encoding has the potential to remove depth ambiguities and provides better depth estimates compared to state-of-the-art approaches. Additionally, we build a low-cost prototype of our CCA using a photographic film and validate the proposed approach in real scenarios. ### Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across Different Shutter Types - **Authors:** Yue Hu, Gourav Datta, Kira Beerel, Peter Beerel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08136 - **Pdf link:** https://arxiv.org/pdf/2309.08136 - **Abstract** Computer vision (CV) pipelines are typically evaluated on datasets processed by image signal processing (ISP) pipelines even though, for resource-constrained applications, an important research goal is to avoid as many ISP steps as possible. In particular, most CV datasets consist of global shutter (GS) images even though most cameras today use a rolling shutter (RS). This paper studies the impact of different shutter mechanisms on machine learning (ML) object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of Unreal Engine 5 (UE5). In particular, we train and evaluate mainstream detection models with our synthetically-generated paired GS and RS datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities, especially when capturing low-speed objects (e.g., pedestrians). The results of this emulation framework indicate the performance between them are remarkably congruent for coarse-grained detection (mean average precision (mAP) for IOU=0.5), but have significant differences for fine-grained measures of detection accuracy (mAP for IOU=0.5:0.95). This implies that ML pipelines might not need explicit correction for RS for many object detection applications, but mitigating RS effects in ISP-less ML pipelines that target fine-grained location of the objects may need additional research. ### Multi-Scale Estimation for Omni-Directional Saliency Maps Using Learnable Equator Bias - **Authors:** Takao Yamanaka, Tatsuya Suzuki, Taiki Nobutsune, Chenjunlin Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08139 - **Pdf link:** https://arxiv.org/pdf/2309.08139 - **Abstract** Omni-directional images have been used in wide range of applications. For the applications, it would be useful to estimate saliency maps representing probability distributions of gazing points with a head-mounted display, to detect important regions in the omni-directional images. This paper proposes a novel saliency-map estimation model for the omni-directional images by extracting overlapping 2-dimensional (2D) plane images from omni-directional images at various directions and angles of view. While 2D saliency maps tend to have high probability at the center of images (center bias), the high-probability region appears at horizontal directions in omni-directional saliency maps when a head-mounted display is used (equator bias). Therefore, the 2D saliency model with a center-bias layer was fine-tuned with an omni-directional dataset by replacing the center-bias layer to an equator-bias layer conditioned on the elevation angle for the extraction of the 2D plane image. The limited availability of omni-directional images in saliency datasets can be compensated by using the well-established 2D saliency model pretrained by a large number of training images with the ground truth of 2D saliency maps. In addition, this paper proposes a multi-scale estimation method by extracting 2D images in multiple angles of view to detect objects of various sizes with variable receptive fields. The saliency maps estimated from the multiple angles of view were integrated by using pixel-wise attention weights calculated in an integration layer for weighting the optimal scale to each object. The proposed method was evaluated using a publicly available dataset with evaluation metrics for omni-directional saliency maps. It was confirmed that the accuracy of the saliency maps was improved by the proposed method. ### AdSEE: Investigating the Impact of Image Style Editing on Advertisement Attractiveness - **Authors:** Liyao Jiang, Chenglin Li, Haolan Chen, Xiaodong Gao, Xinwang Zhong, Yang Qiu, Shani Ye, Di Niu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Information Retrieval (cs.IR); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2309.08159 - **Pdf link:** https://arxiv.org/pdf/2309.08159 - **Abstract** Online advertisements are important elements in e-commerce sites, social media platforms, and search engines. With the increasing popularity of mobile browsing, many online ads are displayed with visual information in the form of a cover image in addition to text descriptions to grab the attention of users. Various recent studies have focused on predicting the click rates of online advertisements aware of visual features or composing optimal advertisement elements to enhance visibility. In this paper, we propose Advertisement Style Editing and Attractiveness Enhancement (AdSEE), which explores whether semantic editing to ads images can affect or alter the popularity of online advertisements. We introduce StyleGAN-based facial semantic editing and inversion to ads images and train a click rate predictor attributing GAN-based face latent representations in addition to traditional visual and textual features to click rates. Through a large collected dataset named QQ-AD, containing 20,527 online ads, we perform extensive offline tests to study how different semantic directions and their edit coefficients may impact click rates. We further design a Genetic Advertisement Editor to efficiently search for the optimal edit directions and intensity given an input ad cover image to enhance its projected click rates. Online A/B tests performed over a period of 5 days have verified the increased click-through rates of AdSEE-edited samples as compared to a control group of original ads, verifying the relation between image styles and ad popularity. We open source the code for AdSEE research at https://github.com/LiyaoJiang1998/adsee. ### Toward responsible face datasets: modeling the distribution of a disentangled latent space for sampling face images from demographic groups - **Authors:** Parsa Rahimi, Christophe Ecabert, Sebastien Marcel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2309.08442 - **Pdf link:** https://arxiv.org/pdf/2309.08442 - **Abstract** Recently, it has been exposed that some modern facial recognition systems could discriminate specific demographic groups and may lead to unfair attention with respect to various facial attributes such as gender and origin. The main reason are the biases inside datasets, unbalanced demographics, used to train theses models. Unfortunately, collecting a large-scale balanced dataset with respect to various demographics is impracticable. In this paper, we investigate as an alternative the generation of a balanced and possibly bias-free synthetic dataset that could be used to train, to regularize or to evaluate deep learning-based facial recognition models. We propose to use a simple method for modeling and sampling a disentangled projection of a StyleGAN latent space to generate any combination of demographic groups (e.g. $hispanic-female$). Our experiments show that we can synthesis any combination of demographic groups effectively and the identities are different from the original training dataset. We also released the source code. ### Visual Speech Recognition for Low-resource Languages with Automatic Labels From Whisper Model - **Authors:** Jeong Hun Yeo, Minsu Kim, Shinji Watanabe, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2309.08535 - **Pdf link:** https://arxiv.org/pdf/2309.08535 - **Abstract** This paper proposes a powerful Visual Speech Recognition (VSR) method for multiple languages, especially for low-resource languages that have a limited number of labeled data. Different from previous methods that tried to improve the VSR performance for the target language by using knowledge learned from other languages, we explore whether we can increase the amount of training data itself for the different languages without human intervention. To this end, we employ a Whisper model which can conduct both language identification and audio-based speech recognition. It serves to filter data of the desired languages and transcribe labels from the unannotated, multilingual audio-visual data pool. By comparing the performances of VSR models trained on automatic labels and the human-annotated labels, we show that we can achieve similar VSR performance to that of human-annotated labels even without utilizing human annotations. Through the automated labeling process, we label large-scale unlabeled multilingual databases, VoxCeleb2 and AVSpeech, producing 1,002 hours of data for four low VSR resource languages, French, Italian, Spanish, and Portuguese. With the automatic labels, we achieve new state-of-the-art performance on mTEDx in four languages, significantly surpassing the previous methods. The automatic labels are available online: https://github.com/JeongHun0716/Visual-Speech-Recognition-for-Low-Resource-Languages ## Keyword: image signal processing ### Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across Different Shutter Types - **Authors:** Yue Hu, Gourav Datta, Kira Beerel, Peter Beerel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08136 - **Pdf link:** https://arxiv.org/pdf/2309.08136 - **Abstract** Computer vision (CV) pipelines are typically evaluated on datasets processed by image signal processing (ISP) pipelines even though, for resource-constrained applications, an important research goal is to avoid as many ISP steps as possible. In particular, most CV datasets consist of global shutter (GS) images even though most cameras today use a rolling shutter (RS). This paper studies the impact of different shutter mechanisms on machine learning (ML) object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of Unreal Engine 5 (UE5). In particular, we train and evaluate mainstream detection models with our synthetically-generated paired GS and RS datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities, especially when capturing low-speed objects (e.g., pedestrians). The results of this emulation framework indicate the performance between them are remarkably congruent for coarse-grained detection (mean average precision (mAP) for IOU=0.5), but have significant differences for fine-grained measures of detection accuracy (mAP for IOU=0.5:0.95). This implies that ML pipelines might not need explicit correction for RS for many object detection applications, but mitigating RS effects in ISP-less ML pipelines that target fine-grained location of the objects may need additional research. ## Keyword: image signal process ### Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across Different Shutter Types - **Authors:** Yue Hu, Gourav Datta, Kira Beerel, Peter Beerel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08136 - **Pdf link:** https://arxiv.org/pdf/2309.08136 - **Abstract** Computer vision (CV) pipelines are typically evaluated on datasets processed by image signal processing (ISP) pipelines even though, for resource-constrained applications, an important research goal is to avoid as many ISP steps as possible. In particular, most CV datasets consist of global shutter (GS) images even though most cameras today use a rolling shutter (RS). This paper studies the impact of different shutter mechanisms on machine learning (ML) object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of Unreal Engine 5 (UE5). In particular, we train and evaluate mainstream detection models with our synthetically-generated paired GS and RS datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities, especially when capturing low-speed objects (e.g., pedestrians). The results of this emulation framework indicate the performance between them are remarkably congruent for coarse-grained detection (mean average precision (mAP) for IOU=0.5), but have significant differences for fine-grained measures of detection accuracy (mAP for IOU=0.5:0.95). This implies that ML pipelines might not need explicit correction for RS for many object detection applications, but mitigating RS effects in ISP-less ML pipelines that target fine-grained location of the objects may need additional research. ## Keyword: compression ### Differentiable Resolution Compression and Alignment for Efficient Video Classification and Retrieval - **Authors:** Rui Deng, Qian Wu, Yuke Li, Haoran Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08167 - **Pdf link:** https://arxiv.org/pdf/2309.08167 - **Abstract** Optimizing video inference efficiency has become increasingly important with the growing demand for video analysis in various fields. Some existing methods achieve high efficiency by explicit discard of spatial or temporal information, which poses challenges in fast-changing and fine-grained scenarios. To address these issues, we propose an efficient video representation network with Differentiable Resolution Compression and Alignment mechanism, which compresses non-essential information in the early stage of the network to reduce computational costs while maintaining consistent temporal correlations. Specifically, we leverage a Differentiable Context-aware Compression Module to encode the saliency and non-saliency frame features, refining and updating the features into a high-low resolution video sequence. To process the new sequence, we introduce a new Resolution-Align Transformer Layer to capture global temporal correlations among frame features with different resolutions, while reducing spatial computation costs quadratically by utilizing fewer spatial tokens in low-resolution non-saliency frames. The entire network can be end-to-end optimized via the integration of the differentiable compression module. Experimental results show that our method achieves the best trade-off between efficiency and performance on near-duplicate video retrieval and competitive results on dynamic video classification compared to state-of-the-art methods. Code:https://github.com/dun-research/DRCA ## Keyword: RAW ### Interpretability-Aware Vision Transformer - **Authors:** Yao Qiang, Chengyin Li, Prashant Khanduri, Dongxiao Zhu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08035 - **Pdf link:** https://arxiv.org/pdf/2309.08035 - **Abstract** Vision Transformers (ViTs) have become prominent models for solving various vision tasks. However, the interpretability of ViTs has not kept pace with their promising performance. While there has been a surge of interest in developing {\it post hoc} solutions to explain ViTs' outputs, these methods do not generalize to different downstream tasks and various transformer architectures. Furthermore, if ViTs are not properly trained with the given data and do not prioritize the region of interest, the {\it post hoc} methods would be less effective. Instead of developing another {\it post hoc} approach, we introduce a novel training procedure that inherently enhances model interpretability. Our interpretability-aware ViT (IA-ViT) draws inspiration from a fresh insight: both the class patch and image patches consistently generate predicted distributions and attention maps. IA-ViT is composed of a feature extractor, a predictor, and an interpreter, which are trained jointly with an interpretability-aware training objective. Consequently, the interpreter simulates the behavior of the predictor and provides a faithful explanation through its single-head self-attention mechanism. Our comprehensive experimental results demonstrate the effectiveness of IA-ViT in several image classification tasks, with both qualitative and quantitative evaluations of model performance and interpretability. Source code is available from: https://github.com/qiangyao1988/IA-ViT. ### Salient Object Detection in Optical Remote Sensing Images Driven by Transformer - **Authors:** Gongyang Li, Zhen Bai, Zhi Liu, Xinpeng Zhang, Haibin Ling - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08206 - **Pdf link:** https://arxiv.org/pdf/2309.08206 - **Abstract** Existing methods for Salient Object Detection in Optical Remote Sensing Images (ORSI-SOD) mainly adopt Convolutional Neural Networks (CNNs) as the backbone, such as VGG and ResNet. Since CNNs can only extract features within certain receptive fields, most ORSI-SOD methods generally follow the local-to-contextual paradigm. In this paper, we propose a novel Global Extraction Local Exploration Network (GeleNet) for ORSI-SOD following the global-to-local paradigm. Specifically, GeleNet first adopts a transformer backbone to generate four-level feature embeddings with global long-range dependencies. Then, GeleNet employs a Direction-aware Shuffle Weighted Spatial Attention Module (D-SWSAM) and its simplified version (SWSAM) to enhance local interactions, and a Knowledge Transfer Module (KTM) to further enhance cross-level contextual interactions. D-SWSAM comprehensively perceives the orientation information in the lowest-level features through directional convolutions to adapt to various orientations of salient objects in ORSIs, and effectively enhances the details of salient objects with an improved attention mechanism. SWSAM discards the direction-aware part of D-SWSAM to focus on localizing salient objects in the highest-level features. KTM models the contextual correlation knowledge of two middle-level features of different scales based on the self-attention mechanism, and transfers the knowledge to the raw features to generate more discriminative features. Finally, a saliency predictor is used to generate the saliency map based on the outputs of the above three modules. Extensive experiments on three public datasets demonstrate that the proposed GeleNet outperforms relevant state-of-the-art methods. The code and results of our method are available at https://github.com/MathLee/GeleNet. ### Beyond Domain Gap: Exploiting Subjectivity in Sketch-Based Person Retrieval - **Authors:** Kejun Lin, Zhixiang Wang, Zheng Wang, Yinqiang Zheng, Shin'ichi Satoh - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM) - **Arxiv link:** https://arxiv.org/abs/2309.08372 - **Pdf link:** https://arxiv.org/pdf/2309.08372 - **Abstract** Person re-identification (re-ID) requires densely distributed cameras. In practice, the person of interest may not be captured by cameras and, therefore, needs to be retrieved using subjective information (e.g., sketches from witnesses). Previous research defines this case using the sketch as sketch re-identification (Sketch re-ID) and focuses on eliminating the domain gap. Actually, subjectivity is another significant challenge. We model and investigate it by posing a new dataset with multi-witness descriptions. It features two aspects. 1) Large-scale. It contains over 4,763 sketches and 32,668 photos, making it the largest Sketch re-ID dataset. 2) Multi-perspective and multi-style. Our dataset offers multiple sketches for each identity. Witnesses' subjective cognition provides multiple perspectives on the same individual, while different artists' drawing styles provide variation in sketch styles. We further have two novel designs to alleviate the challenge of subjectivity. 1) Fusing subjectivity. We propose a non-local (NL) fusion module that gathers sketches from different witnesses for the same identity. 2) Introducing objectivity. An AttrAlign module utilizes attributes as an implicit mask to align cross-domain features. To push forward the advance of Sketch re-ID, we set three benchmarks (large-scale, multi-style, cross-style). Extensive experiments demonstrate our leading performance in these benchmarks. Dataset and Codes are publicly available at: https://github.com/Lin-Kayla/subjectivity-sketch-reid ### Towards Practical and Efficient Image-to-Speech Captioning with Vision-Language Pre-training and Multi-modal Tokens - **Authors:** Minsu Kim, Jeongsoo Choi, Soumi Maiti, Jeong Hun Yeo, Shinji Watanabe, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL); Audio and Speech Processing (eess.AS); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2309.08531 - **Pdf link:** https://arxiv.org/pdf/2309.08531 - **Abstract** In this paper, we propose methods to build a powerful and efficient Image-to-Speech captioning (Im2Sp) model. To this end, we start with importing the rich knowledge related to image comprehension and language modeling from a large-scale pre-trained vision-language model into Im2Sp. We set the output of the proposed Im2Sp as discretized speech units, i.e., the quantized speech features of a self-supervised speech model. The speech units mainly contain linguistic information while suppressing other characteristics of speech. This allows us to incorporate the language modeling capability of the pre-trained vision-language model into the spoken language modeling of Im2Sp. With the vision-language pre-training strategy, we set new state-of-the-art Im2Sp performances on two widely used benchmark databases, COCO and Flickr8k. Then, we further improve the efficiency of the Im2Sp model. Similar to the speech unit case, we convert the original image into image units, which are derived through vector quantization of the raw image. With these image units, we can drastically reduce the required data storage for saving image data to just 0.8% when compared to the original image data in terms of bits. Demo page: https://ms-dot-k.github.io/Image-to-Speech-Captioning. ## Keyword: raw image ### Towards Practical and Efficient Image-to-Speech Captioning with Vision-Language Pre-training and Multi-modal Tokens - **Authors:** Minsu Kim, Jeongsoo Choi, Soumi Maiti, Jeong Hun Yeo, Shinji Watanabe, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL); Audio and Speech Processing (eess.AS); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2309.08531 - **Pdf link:** https://arxiv.org/pdf/2309.08531 - **Abstract** In this paper, we propose methods to build a powerful and efficient Image-to-Speech captioning (Im2Sp) model. To this end, we start with importing the rich knowledge related to image comprehension and language modeling from a large-scale pre-trained vision-language model into Im2Sp. We set the output of the proposed Im2Sp as discretized speech units, i.e., the quantized speech features of a self-supervised speech model. The speech units mainly contain linguistic information while suppressing other characteristics of speech. This allows us to incorporate the language modeling capability of the pre-trained vision-language model into the spoken language modeling of Im2Sp. With the vision-language pre-training strategy, we set new state-of-the-art Im2Sp performances on two widely used benchmark databases, COCO and Flickr8k. Then, we further improve the efficiency of the Im2Sp model. Similar to the speech unit case, we convert the original image into image units, which are derived through vector quantization of the raw image. With these image units, we can drastically reduce the required data storage for saving image data to just 0.8% when compared to the original image data in terms of bits. Demo page: https://ms-dot-k.github.io/Image-to-Speech-Captioning.
2.0
New submissions for Mon, 18 Sep 23 - ## Keyword: events ### Deformable Neural Radiance Fields using RGB and Event Cameras - **Authors:** Qi Ma, Danda Pani Paudel, Ajad Chhatkuli, Luc Van Gool - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08416 - **Pdf link:** https://arxiv.org/pdf/2309.08416 - **Abstract** Modeling Neural Radiance Fields for fast-moving deformable objects from visual data alone is a challenging problem. A major issue arises due to the high deformation and low acquisition rates. To address this problem, we propose to use event cameras that offer very fast acquisition of visual change in an asynchronous manner. In this work, we develop a novel method to model the deformable neural radiance fields using RGB and event cameras. The proposed method uses the asynchronous stream of events and calibrated sparse RGB frames. In our setup, the camera pose at the individual events required to integrate them into the radiance fields remains unknown. Our method jointly optimizes these poses and the radiance field. This happens efficiently by leveraging the collection of events at once and actively sampling the events during learning. Experiments conducted on both realistically rendered graphics and real-world datasets demonstrate a significant benefit of the proposed method over the state-of-the-art and the compared baseline. This shows a promising direction for modeling deformable neural radiance fields in real-world dynamic scenes. ### Robust e-NeRF: NeRF from Sparse & Noisy Events under Non-Uniform Motion - **Authors:** Weng Fei Low, Gim Hee Lee - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2309.08596 - **Pdf link:** https://arxiv.org/pdf/2309.08596 - **Abstract** Event cameras offer many advantages over standard cameras due to their distinctive principle of operation: low power, low latency, high temporal resolution and high dynamic range. Nonetheless, the success of many downstream visual applications also hinges on an efficient and effective scene representation, where Neural Radiance Field (NeRF) is seen as the leading candidate. Such promise and potential of event cameras and NeRF inspired recent works to investigate on the reconstruction of NeRF from moving event cameras. However, these works are mainly limited in terms of the dependence on dense and low-noise event streams, as well as generalization to arbitrary contrast threshold values and camera speed profiles. In this work, we propose Robust e-NeRF, a novel method to directly and robustly reconstruct NeRFs from moving event cameras under various real-world conditions, especially from sparse and noisy events generated under non-uniform motion. It consists of two key components: a realistic event generation model that accounts for various intrinsic parameters (e.g. time-independent, asymmetric threshold and refractory period) and non-idealities (e.g. pixel-to-pixel threshold variation), as well as a complementary pair of normalized reconstruction losses that can effectively generalize to arbitrary speed profiles and intrinsic parameter values without such prior knowledge. Experiments on real and novel realistically simulated sequences verify our effectiveness. Our code, synthetic dataset and improved event simulator are public. ## Keyword: event camera ### Deformable Neural Radiance Fields using RGB and Event Cameras - **Authors:** Qi Ma, Danda Pani Paudel, Ajad Chhatkuli, Luc Van Gool - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08416 - **Pdf link:** https://arxiv.org/pdf/2309.08416 - **Abstract** Modeling Neural Radiance Fields for fast-moving deformable objects from visual data alone is a challenging problem. A major issue arises due to the high deformation and low acquisition rates. To address this problem, we propose to use event cameras that offer very fast acquisition of visual change in an asynchronous manner. In this work, we develop a novel method to model the deformable neural radiance fields using RGB and event cameras. The proposed method uses the asynchronous stream of events and calibrated sparse RGB frames. In our setup, the camera pose at the individual events required to integrate them into the radiance fields remains unknown. Our method jointly optimizes these poses and the radiance field. This happens efficiently by leveraging the collection of events at once and actively sampling the events during learning. Experiments conducted on both realistically rendered graphics and real-world datasets demonstrate a significant benefit of the proposed method over the state-of-the-art and the compared baseline. This shows a promising direction for modeling deformable neural radiance fields in real-world dynamic scenes. ### Robust e-NeRF: NeRF from Sparse & Noisy Events under Non-Uniform Motion - **Authors:** Weng Fei Low, Gim Hee Lee - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2309.08596 - **Pdf link:** https://arxiv.org/pdf/2309.08596 - **Abstract** Event cameras offer many advantages over standard cameras due to their distinctive principle of operation: low power, low latency, high temporal resolution and high dynamic range. Nonetheless, the success of many downstream visual applications also hinges on an efficient and effective scene representation, where Neural Radiance Field (NeRF) is seen as the leading candidate. Such promise and potential of event cameras and NeRF inspired recent works to investigate on the reconstruction of NeRF from moving event cameras. However, these works are mainly limited in terms of the dependence on dense and low-noise event streams, as well as generalization to arbitrary contrast threshold values and camera speed profiles. In this work, we propose Robust e-NeRF, a novel method to directly and robustly reconstruct NeRFs from moving event cameras under various real-world conditions, especially from sparse and noisy events generated under non-uniform motion. It consists of two key components: a realistic event generation model that accounts for various intrinsic parameters (e.g. time-independent, asymmetric threshold and refractory period) and non-idealities (e.g. pixel-to-pixel threshold variation), as well as a complementary pair of normalized reconstruction losses that can effectively generalize to arbitrary speed profiles and intrinsic parameter values without such prior knowledge. Experiments on real and novel realistically simulated sequences verify our effectiveness. Our code, synthetic dataset and improved event simulator are public. ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Depth Estimation from a Single Optical Encoded Image using a Learned Colored-Coded Aperture - **Authors:** Jhon Lopez, Edwin Vargas, Henry Arguello - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08033 - **Pdf link:** https://arxiv.org/pdf/2309.08033 - **Abstract** Depth estimation from a single image of a conventional camera is a challenging task since depth cues are lost during the acquisition process. State-of-the-art approaches improve the discrimination between different depths by introducing a binary-coded aperture (CA) in the lens aperture that generates different coded blur patterns at different depths. Color-coded apertures (CCA) can also produce color misalignment in the captured image which can be utilized to estimate disparity. Leveraging advances in deep learning, more recent works have explored the data-driven design of a diffractive optical element (DOE) for encoding depth information through chromatic aberrations. However, compared with binary CA or CCA, DOEs are more expensive to fabricate and require high-precision devices. Different from previous CCA-based approaches that employ few basic colors, in this work we propose a CCA with a greater number of color filters and richer spectral information to optically encode relevant depth information in a single snapshot. Furthermore, we propose to jointly learn the color-coded aperture (CCA) pattern and a convolutional neural network (CNN) to retrieve depth information by using an end-to-end optimization approach. We demonstrate through different experiments on three different data sets that the designed color-encoding has the potential to remove depth ambiguities and provides better depth estimates compared to state-of-the-art approaches. Additionally, we build a low-cost prototype of our CCA using a photographic film and validate the proposed approach in real scenarios. ### Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across Different Shutter Types - **Authors:** Yue Hu, Gourav Datta, Kira Beerel, Peter Beerel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08136 - **Pdf link:** https://arxiv.org/pdf/2309.08136 - **Abstract** Computer vision (CV) pipelines are typically evaluated on datasets processed by image signal processing (ISP) pipelines even though, for resource-constrained applications, an important research goal is to avoid as many ISP steps as possible. In particular, most CV datasets consist of global shutter (GS) images even though most cameras today use a rolling shutter (RS). This paper studies the impact of different shutter mechanisms on machine learning (ML) object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of Unreal Engine 5 (UE5). In particular, we train and evaluate mainstream detection models with our synthetically-generated paired GS and RS datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities, especially when capturing low-speed objects (e.g., pedestrians). The results of this emulation framework indicate the performance between them are remarkably congruent for coarse-grained detection (mean average precision (mAP) for IOU=0.5), but have significant differences for fine-grained measures of detection accuracy (mAP for IOU=0.5:0.95). This implies that ML pipelines might not need explicit correction for RS for many object detection applications, but mitigating RS effects in ISP-less ML pipelines that target fine-grained location of the objects may need additional research. ### Multi-Scale Estimation for Omni-Directional Saliency Maps Using Learnable Equator Bias - **Authors:** Takao Yamanaka, Tatsuya Suzuki, Taiki Nobutsune, Chenjunlin Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08139 - **Pdf link:** https://arxiv.org/pdf/2309.08139 - **Abstract** Omni-directional images have been used in wide range of applications. For the applications, it would be useful to estimate saliency maps representing probability distributions of gazing points with a head-mounted display, to detect important regions in the omni-directional images. This paper proposes a novel saliency-map estimation model for the omni-directional images by extracting overlapping 2-dimensional (2D) plane images from omni-directional images at various directions and angles of view. While 2D saliency maps tend to have high probability at the center of images (center bias), the high-probability region appears at horizontal directions in omni-directional saliency maps when a head-mounted display is used (equator bias). Therefore, the 2D saliency model with a center-bias layer was fine-tuned with an omni-directional dataset by replacing the center-bias layer to an equator-bias layer conditioned on the elevation angle for the extraction of the 2D plane image. The limited availability of omni-directional images in saliency datasets can be compensated by using the well-established 2D saliency model pretrained by a large number of training images with the ground truth of 2D saliency maps. In addition, this paper proposes a multi-scale estimation method by extracting 2D images in multiple angles of view to detect objects of various sizes with variable receptive fields. The saliency maps estimated from the multiple angles of view were integrated by using pixel-wise attention weights calculated in an integration layer for weighting the optimal scale to each object. The proposed method was evaluated using a publicly available dataset with evaluation metrics for omni-directional saliency maps. It was confirmed that the accuracy of the saliency maps was improved by the proposed method. ### AdSEE: Investigating the Impact of Image Style Editing on Advertisement Attractiveness - **Authors:** Liyao Jiang, Chenglin Li, Haolan Chen, Xiaodong Gao, Xinwang Zhong, Yang Qiu, Shani Ye, Di Niu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Information Retrieval (cs.IR); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2309.08159 - **Pdf link:** https://arxiv.org/pdf/2309.08159 - **Abstract** Online advertisements are important elements in e-commerce sites, social media platforms, and search engines. With the increasing popularity of mobile browsing, many online ads are displayed with visual information in the form of a cover image in addition to text descriptions to grab the attention of users. Various recent studies have focused on predicting the click rates of online advertisements aware of visual features or composing optimal advertisement elements to enhance visibility. In this paper, we propose Advertisement Style Editing and Attractiveness Enhancement (AdSEE), which explores whether semantic editing to ads images can affect or alter the popularity of online advertisements. We introduce StyleGAN-based facial semantic editing and inversion to ads images and train a click rate predictor attributing GAN-based face latent representations in addition to traditional visual and textual features to click rates. Through a large collected dataset named QQ-AD, containing 20,527 online ads, we perform extensive offline tests to study how different semantic directions and their edit coefficients may impact click rates. We further design a Genetic Advertisement Editor to efficiently search for the optimal edit directions and intensity given an input ad cover image to enhance its projected click rates. Online A/B tests performed over a period of 5 days have verified the increased click-through rates of AdSEE-edited samples as compared to a control group of original ads, verifying the relation between image styles and ad popularity. We open source the code for AdSEE research at https://github.com/LiyaoJiang1998/adsee. ### Toward responsible face datasets: modeling the distribution of a disentangled latent space for sampling face images from demographic groups - **Authors:** Parsa Rahimi, Christophe Ecabert, Sebastien Marcel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2309.08442 - **Pdf link:** https://arxiv.org/pdf/2309.08442 - **Abstract** Recently, it has been exposed that some modern facial recognition systems could discriminate specific demographic groups and may lead to unfair attention with respect to various facial attributes such as gender and origin. The main reason are the biases inside datasets, unbalanced demographics, used to train theses models. Unfortunately, collecting a large-scale balanced dataset with respect to various demographics is impracticable. In this paper, we investigate as an alternative the generation of a balanced and possibly bias-free synthetic dataset that could be used to train, to regularize or to evaluate deep learning-based facial recognition models. We propose to use a simple method for modeling and sampling a disentangled projection of a StyleGAN latent space to generate any combination of demographic groups (e.g. $hispanic-female$). Our experiments show that we can synthesis any combination of demographic groups effectively and the identities are different from the original training dataset. We also released the source code. ### Visual Speech Recognition for Low-resource Languages with Automatic Labels From Whisper Model - **Authors:** Jeong Hun Yeo, Minsu Kim, Shinji Watanabe, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2309.08535 - **Pdf link:** https://arxiv.org/pdf/2309.08535 - **Abstract** This paper proposes a powerful Visual Speech Recognition (VSR) method for multiple languages, especially for low-resource languages that have a limited number of labeled data. Different from previous methods that tried to improve the VSR performance for the target language by using knowledge learned from other languages, we explore whether we can increase the amount of training data itself for the different languages without human intervention. To this end, we employ a Whisper model which can conduct both language identification and audio-based speech recognition. It serves to filter data of the desired languages and transcribe labels from the unannotated, multilingual audio-visual data pool. By comparing the performances of VSR models trained on automatic labels and the human-annotated labels, we show that we can achieve similar VSR performance to that of human-annotated labels even without utilizing human annotations. Through the automated labeling process, we label large-scale unlabeled multilingual databases, VoxCeleb2 and AVSpeech, producing 1,002 hours of data for four low VSR resource languages, French, Italian, Spanish, and Portuguese. With the automatic labels, we achieve new state-of-the-art performance on mTEDx in four languages, significantly surpassing the previous methods. The automatic labels are available online: https://github.com/JeongHun0716/Visual-Speech-Recognition-for-Low-Resource-Languages ## Keyword: image signal processing ### Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across Different Shutter Types - **Authors:** Yue Hu, Gourav Datta, Kira Beerel, Peter Beerel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08136 - **Pdf link:** https://arxiv.org/pdf/2309.08136 - **Abstract** Computer vision (CV) pipelines are typically evaluated on datasets processed by image signal processing (ISP) pipelines even though, for resource-constrained applications, an important research goal is to avoid as many ISP steps as possible. In particular, most CV datasets consist of global shutter (GS) images even though most cameras today use a rolling shutter (RS). This paper studies the impact of different shutter mechanisms on machine learning (ML) object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of Unreal Engine 5 (UE5). In particular, we train and evaluate mainstream detection models with our synthetically-generated paired GS and RS datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities, especially when capturing low-speed objects (e.g., pedestrians). The results of this emulation framework indicate the performance between them are remarkably congruent for coarse-grained detection (mean average precision (mAP) for IOU=0.5), but have significant differences for fine-grained measures of detection accuracy (mAP for IOU=0.5:0.95). This implies that ML pipelines might not need explicit correction for RS for many object detection applications, but mitigating RS effects in ISP-less ML pipelines that target fine-grained location of the objects may need additional research. ## Keyword: image signal process ### Let's Roll: Synthetic Dataset Analysis for Pedestrian Detection Across Different Shutter Types - **Authors:** Yue Hu, Gourav Datta, Kira Beerel, Peter Beerel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08136 - **Pdf link:** https://arxiv.org/pdf/2309.08136 - **Abstract** Computer vision (CV) pipelines are typically evaluated on datasets processed by image signal processing (ISP) pipelines even though, for resource-constrained applications, an important research goal is to avoid as many ISP steps as possible. In particular, most CV datasets consist of global shutter (GS) images even though most cameras today use a rolling shutter (RS). This paper studies the impact of different shutter mechanisms on machine learning (ML) object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of Unreal Engine 5 (UE5). In particular, we train and evaluate mainstream detection models with our synthetically-generated paired GS and RS datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities, especially when capturing low-speed objects (e.g., pedestrians). The results of this emulation framework indicate the performance between them are remarkably congruent for coarse-grained detection (mean average precision (mAP) for IOU=0.5), but have significant differences for fine-grained measures of detection accuracy (mAP for IOU=0.5:0.95). This implies that ML pipelines might not need explicit correction for RS for many object detection applications, but mitigating RS effects in ISP-less ML pipelines that target fine-grained location of the objects may need additional research. ## Keyword: compression ### Differentiable Resolution Compression and Alignment for Efficient Video Classification and Retrieval - **Authors:** Rui Deng, Qian Wu, Yuke Li, Haoran Fu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08167 - **Pdf link:** https://arxiv.org/pdf/2309.08167 - **Abstract** Optimizing video inference efficiency has become increasingly important with the growing demand for video analysis in various fields. Some existing methods achieve high efficiency by explicit discard of spatial or temporal information, which poses challenges in fast-changing and fine-grained scenarios. To address these issues, we propose an efficient video representation network with Differentiable Resolution Compression and Alignment mechanism, which compresses non-essential information in the early stage of the network to reduce computational costs while maintaining consistent temporal correlations. Specifically, we leverage a Differentiable Context-aware Compression Module to encode the saliency and non-saliency frame features, refining and updating the features into a high-low resolution video sequence. To process the new sequence, we introduce a new Resolution-Align Transformer Layer to capture global temporal correlations among frame features with different resolutions, while reducing spatial computation costs quadratically by utilizing fewer spatial tokens in low-resolution non-saliency frames. The entire network can be end-to-end optimized via the integration of the differentiable compression module. Experimental results show that our method achieves the best trade-off between efficiency and performance on near-duplicate video retrieval and competitive results on dynamic video classification compared to state-of-the-art methods. Code:https://github.com/dun-research/DRCA ## Keyword: RAW ### Interpretability-Aware Vision Transformer - **Authors:** Yao Qiang, Chengyin Li, Prashant Khanduri, Dongxiao Zhu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08035 - **Pdf link:** https://arxiv.org/pdf/2309.08035 - **Abstract** Vision Transformers (ViTs) have become prominent models for solving various vision tasks. However, the interpretability of ViTs has not kept pace with their promising performance. While there has been a surge of interest in developing {\it post hoc} solutions to explain ViTs' outputs, these methods do not generalize to different downstream tasks and various transformer architectures. Furthermore, if ViTs are not properly trained with the given data and do not prioritize the region of interest, the {\it post hoc} methods would be less effective. Instead of developing another {\it post hoc} approach, we introduce a novel training procedure that inherently enhances model interpretability. Our interpretability-aware ViT (IA-ViT) draws inspiration from a fresh insight: both the class patch and image patches consistently generate predicted distributions and attention maps. IA-ViT is composed of a feature extractor, a predictor, and an interpreter, which are trained jointly with an interpretability-aware training objective. Consequently, the interpreter simulates the behavior of the predictor and provides a faithful explanation through its single-head self-attention mechanism. Our comprehensive experimental results demonstrate the effectiveness of IA-ViT in several image classification tasks, with both qualitative and quantitative evaluations of model performance and interpretability. Source code is available from: https://github.com/qiangyao1988/IA-ViT. ### Salient Object Detection in Optical Remote Sensing Images Driven by Transformer - **Authors:** Gongyang Li, Zhen Bai, Zhi Liu, Xinpeng Zhang, Haibin Ling - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2309.08206 - **Pdf link:** https://arxiv.org/pdf/2309.08206 - **Abstract** Existing methods for Salient Object Detection in Optical Remote Sensing Images (ORSI-SOD) mainly adopt Convolutional Neural Networks (CNNs) as the backbone, such as VGG and ResNet. Since CNNs can only extract features within certain receptive fields, most ORSI-SOD methods generally follow the local-to-contextual paradigm. In this paper, we propose a novel Global Extraction Local Exploration Network (GeleNet) for ORSI-SOD following the global-to-local paradigm. Specifically, GeleNet first adopts a transformer backbone to generate four-level feature embeddings with global long-range dependencies. Then, GeleNet employs a Direction-aware Shuffle Weighted Spatial Attention Module (D-SWSAM) and its simplified version (SWSAM) to enhance local interactions, and a Knowledge Transfer Module (KTM) to further enhance cross-level contextual interactions. D-SWSAM comprehensively perceives the orientation information in the lowest-level features through directional convolutions to adapt to various orientations of salient objects in ORSIs, and effectively enhances the details of salient objects with an improved attention mechanism. SWSAM discards the direction-aware part of D-SWSAM to focus on localizing salient objects in the highest-level features. KTM models the contextual correlation knowledge of two middle-level features of different scales based on the self-attention mechanism, and transfers the knowledge to the raw features to generate more discriminative features. Finally, a saliency predictor is used to generate the saliency map based on the outputs of the above three modules. Extensive experiments on three public datasets demonstrate that the proposed GeleNet outperforms relevant state-of-the-art methods. The code and results of our method are available at https://github.com/MathLee/GeleNet. ### Beyond Domain Gap: Exploiting Subjectivity in Sketch-Based Person Retrieval - **Authors:** Kejun Lin, Zhixiang Wang, Zheng Wang, Yinqiang Zheng, Shin'ichi Satoh - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM) - **Arxiv link:** https://arxiv.org/abs/2309.08372 - **Pdf link:** https://arxiv.org/pdf/2309.08372 - **Abstract** Person re-identification (re-ID) requires densely distributed cameras. In practice, the person of interest may not be captured by cameras and, therefore, needs to be retrieved using subjective information (e.g., sketches from witnesses). Previous research defines this case using the sketch as sketch re-identification (Sketch re-ID) and focuses on eliminating the domain gap. Actually, subjectivity is another significant challenge. We model and investigate it by posing a new dataset with multi-witness descriptions. It features two aspects. 1) Large-scale. It contains over 4,763 sketches and 32,668 photos, making it the largest Sketch re-ID dataset. 2) Multi-perspective and multi-style. Our dataset offers multiple sketches for each identity. Witnesses' subjective cognition provides multiple perspectives on the same individual, while different artists' drawing styles provide variation in sketch styles. We further have two novel designs to alleviate the challenge of subjectivity. 1) Fusing subjectivity. We propose a non-local (NL) fusion module that gathers sketches from different witnesses for the same identity. 2) Introducing objectivity. An AttrAlign module utilizes attributes as an implicit mask to align cross-domain features. To push forward the advance of Sketch re-ID, we set three benchmarks (large-scale, multi-style, cross-style). Extensive experiments demonstrate our leading performance in these benchmarks. Dataset and Codes are publicly available at: https://github.com/Lin-Kayla/subjectivity-sketch-reid ### Towards Practical and Efficient Image-to-Speech Captioning with Vision-Language Pre-training and Multi-modal Tokens - **Authors:** Minsu Kim, Jeongsoo Choi, Soumi Maiti, Jeong Hun Yeo, Shinji Watanabe, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL); Audio and Speech Processing (eess.AS); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2309.08531 - **Pdf link:** https://arxiv.org/pdf/2309.08531 - **Abstract** In this paper, we propose methods to build a powerful and efficient Image-to-Speech captioning (Im2Sp) model. To this end, we start with importing the rich knowledge related to image comprehension and language modeling from a large-scale pre-trained vision-language model into Im2Sp. We set the output of the proposed Im2Sp as discretized speech units, i.e., the quantized speech features of a self-supervised speech model. The speech units mainly contain linguistic information while suppressing other characteristics of speech. This allows us to incorporate the language modeling capability of the pre-trained vision-language model into the spoken language modeling of Im2Sp. With the vision-language pre-training strategy, we set new state-of-the-art Im2Sp performances on two widely used benchmark databases, COCO and Flickr8k. Then, we further improve the efficiency of the Im2Sp model. Similar to the speech unit case, we convert the original image into image units, which are derived through vector quantization of the raw image. With these image units, we can drastically reduce the required data storage for saving image data to just 0.8% when compared to the original image data in terms of bits. Demo page: https://ms-dot-k.github.io/Image-to-Speech-Captioning. ## Keyword: raw image ### Towards Practical and Efficient Image-to-Speech Captioning with Vision-Language Pre-training and Multi-modal Tokens - **Authors:** Minsu Kim, Jeongsoo Choi, Soumi Maiti, Jeong Hun Yeo, Shinji Watanabe, Yong Man Ro - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL); Audio and Speech Processing (eess.AS); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2309.08531 - **Pdf link:** https://arxiv.org/pdf/2309.08531 - **Abstract** In this paper, we propose methods to build a powerful and efficient Image-to-Speech captioning (Im2Sp) model. To this end, we start with importing the rich knowledge related to image comprehension and language modeling from a large-scale pre-trained vision-language model into Im2Sp. We set the output of the proposed Im2Sp as discretized speech units, i.e., the quantized speech features of a self-supervised speech model. The speech units mainly contain linguistic information while suppressing other characteristics of speech. This allows us to incorporate the language modeling capability of the pre-trained vision-language model into the spoken language modeling of Im2Sp. With the vision-language pre-training strategy, we set new state-of-the-art Im2Sp performances on two widely used benchmark databases, COCO and Flickr8k. Then, we further improve the efficiency of the Im2Sp model. Similar to the speech unit case, we convert the original image into image units, which are derived through vector quantization of the raw image. With these image units, we can drastically reduce the required data storage for saving image data to just 0.8% when compared to the original image data in terms of bits. Demo page: https://ms-dot-k.github.io/Image-to-Speech-Captioning.
process
new submissions for mon sep keyword events deformable neural radiance fields using rgb and event cameras authors qi ma danda pani paudel ajad chhatkuli luc van gool subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract modeling neural radiance fields for fast moving deformable objects from visual data alone is a challenging problem a major issue arises due to the high deformation and low acquisition rates to address this problem we propose to use event cameras that offer very fast acquisition of visual change in an asynchronous manner in this work we develop a novel method to model the deformable neural radiance fields using rgb and event cameras the proposed method uses the asynchronous stream of events and calibrated sparse rgb frames in our setup the camera pose at the individual events required to integrate them into the radiance fields remains unknown our method jointly optimizes these poses and the radiance field this happens efficiently by leveraging the collection of events at once and actively sampling the events during learning experiments conducted on both realistically rendered graphics and real world datasets demonstrate a significant benefit of the proposed method over the state of the art and the compared baseline this shows a promising direction for modeling deformable neural radiance fields in real world dynamic scenes robust e nerf nerf from sparse noisy events under non uniform motion authors weng fei low gim hee lee subjects computer vision and pattern recognition cs cv graphics cs gr robotics cs ro arxiv link pdf link abstract event cameras offer many advantages over standard cameras due to their distinctive principle of operation low power low latency high temporal resolution and high dynamic range nonetheless the success of many downstream visual applications also hinges on an efficient and effective scene representation where neural radiance field nerf is seen as the leading candidate such promise and potential of event cameras and nerf inspired recent works to investigate on the reconstruction of nerf from moving event cameras however these works are mainly limited in terms of the dependence on dense and low noise event streams as well as generalization to arbitrary contrast threshold values and camera speed profiles in this work we propose robust e nerf a novel method to directly and robustly reconstruct nerfs from moving event cameras under various real world conditions especially from sparse and noisy events generated under non uniform motion it consists of two key components a realistic event generation model that accounts for various intrinsic parameters e g time independent asymmetric threshold and refractory period and non idealities e g pixel to pixel threshold variation as well as a complementary pair of normalized reconstruction losses that can effectively generalize to arbitrary speed profiles and intrinsic parameter values without such prior knowledge experiments on real and novel realistically simulated sequences verify our effectiveness our code synthetic dataset and improved event simulator are public keyword event camera deformable neural radiance fields using rgb and event cameras authors qi ma danda pani paudel ajad chhatkuli luc van gool subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract modeling neural radiance fields for fast moving deformable objects from visual data alone is a challenging problem a major issue arises due to the high deformation and low acquisition rates to address this problem we propose to use event cameras that offer very fast acquisition of visual change in an asynchronous manner in this work we develop a novel method to model the deformable neural radiance fields using rgb and event cameras the proposed method uses the asynchronous stream of events and calibrated sparse rgb frames in our setup the camera pose at the individual events required to integrate them into the radiance fields remains unknown our method jointly optimizes these poses and the radiance field this happens efficiently by leveraging the collection of events at once and actively sampling the events during learning experiments conducted on both realistically rendered graphics and real world datasets demonstrate a significant benefit of the proposed method over the state of the art and the compared baseline this shows a promising direction for modeling deformable neural radiance fields in real world dynamic scenes robust e nerf nerf from sparse noisy events under non uniform motion authors weng fei low gim hee lee subjects computer vision and pattern recognition cs cv graphics cs gr robotics cs ro arxiv link pdf link abstract event cameras offer many advantages over standard cameras due to their distinctive principle of operation low power low latency high temporal resolution and high dynamic range nonetheless the success of many downstream visual applications also hinges on an efficient and effective scene representation where neural radiance field nerf is seen as the leading candidate such promise and potential of event cameras and nerf inspired recent works to investigate on the reconstruction of nerf from moving event cameras however these works are mainly limited in terms of the dependence on dense and low noise event streams as well as generalization to arbitrary contrast threshold values and camera speed profiles in this work we propose robust e nerf a novel method to directly and robustly reconstruct nerfs from moving event cameras under various real world conditions especially from sparse and noisy events generated under non uniform motion it consists of two key components a realistic event generation model that accounts for various intrinsic parameters e g time independent asymmetric threshold and refractory period and non idealities e g pixel to pixel threshold variation as well as a complementary pair of normalized reconstruction losses that can effectively generalize to arbitrary speed profiles and intrinsic parameter values without such prior knowledge experiments on real and novel realistically simulated sequences verify our effectiveness our code synthetic dataset and improved event simulator are public keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp depth estimation from a single optical encoded image using a learned colored coded aperture authors jhon lopez edwin vargas henry arguello subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract depth estimation from a single image of a conventional camera is a challenging task since depth cues are lost during the acquisition process state of the art approaches improve the discrimination between different depths by introducing a binary coded aperture ca in the lens aperture that generates different coded blur patterns at different depths color coded apertures cca can also produce color misalignment in the captured image which can be utilized to estimate disparity leveraging advances in deep learning more recent works have explored the data driven design of a diffractive optical element doe for encoding depth information through chromatic aberrations however compared with binary ca or cca does are more expensive to fabricate and require high precision devices different from previous cca based approaches that employ few basic colors in this work we propose a cca with a greater number of color filters and richer spectral information to optically encode relevant depth information in a single snapshot furthermore we propose to jointly learn the color coded aperture cca pattern and a convolutional neural network cnn to retrieve depth information by using an end to end optimization approach we demonstrate through different experiments on three different data sets that the designed color encoding has the potential to remove depth ambiguities and provides better depth estimates compared to state of the art approaches additionally we build a low cost prototype of our cca using a photographic film and validate the proposed approach in real scenarios let s roll synthetic dataset analysis for pedestrian detection across different shutter types authors yue hu gourav datta kira beerel peter beerel subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract computer vision cv pipelines are typically evaluated on datasets processed by image signal processing isp pipelines even though for resource constrained applications an important research goal is to avoid as many isp steps as possible in particular most cv datasets consist of global shutter gs images even though most cameras today use a rolling shutter rs this paper studies the impact of different shutter mechanisms on machine learning ml object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of unreal engine in particular we train and evaluate mainstream detection models with our synthetically generated paired gs and rs datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities especially when capturing low speed objects e g pedestrians the results of this emulation framework indicate the performance between them are remarkably congruent for coarse grained detection mean average precision map for iou but have significant differences for fine grained measures of detection accuracy map for iou this implies that ml pipelines might not need explicit correction for rs for many object detection applications but mitigating rs effects in isp less ml pipelines that target fine grained location of the objects may need additional research multi scale estimation for omni directional saliency maps using learnable equator bias authors takao yamanaka tatsuya suzuki taiki nobutsune chenjunlin wu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract omni directional images have been used in wide range of applications for the applications it would be useful to estimate saliency maps representing probability distributions of gazing points with a head mounted display to detect important regions in the omni directional images this paper proposes a novel saliency map estimation model for the omni directional images by extracting overlapping dimensional plane images from omni directional images at various directions and angles of view while saliency maps tend to have high probability at the center of images center bias the high probability region appears at horizontal directions in omni directional saliency maps when a head mounted display is used equator bias therefore the saliency model with a center bias layer was fine tuned with an omni directional dataset by replacing the center bias layer to an equator bias layer conditioned on the elevation angle for the extraction of the plane image the limited availability of omni directional images in saliency datasets can be compensated by using the well established saliency model pretrained by a large number of training images with the ground truth of saliency maps in addition this paper proposes a multi scale estimation method by extracting images in multiple angles of view to detect objects of various sizes with variable receptive fields the saliency maps estimated from the multiple angles of view were integrated by using pixel wise attention weights calculated in an integration layer for weighting the optimal scale to each object the proposed method was evaluated using a publicly available dataset with evaluation metrics for omni directional saliency maps it was confirmed that the accuracy of the saliency maps was improved by the proposed method adsee investigating the impact of image style editing on advertisement attractiveness authors liyao jiang chenglin li haolan chen xiaodong gao xinwang zhong yang qiu shani ye di niu subjects computer vision and pattern recognition cs cv information retrieval cs ir machine learning cs lg arxiv link pdf link abstract online advertisements are important elements in e commerce sites social media platforms and search engines with the increasing popularity of mobile browsing many online ads are displayed with visual information in the form of a cover image in addition to text descriptions to grab the attention of users various recent studies have focused on predicting the click rates of online advertisements aware of visual features or composing optimal advertisement elements to enhance visibility in this paper we propose advertisement style editing and attractiveness enhancement adsee which explores whether semantic editing to ads images can affect or alter the popularity of online advertisements we introduce stylegan based facial semantic editing and inversion to ads images and train a click rate predictor attributing gan based face latent representations in addition to traditional visual and textual features to click rates through a large collected dataset named qq ad containing online ads we perform extensive offline tests to study how different semantic directions and their edit coefficients may impact click rates we further design a genetic advertisement editor to efficiently search for the optimal edit directions and intensity given an input ad cover image to enhance its projected click rates online a b tests performed over a period of days have verified the increased click through rates of adsee edited samples as compared to a control group of original ads verifying the relation between image styles and ad popularity we open source the code for adsee research at toward responsible face datasets modeling the distribution of a disentangled latent space for sampling face images from demographic groups authors parsa rahimi christophe ecabert sebastien marcel subjects computer vision and pattern recognition cs cv artificial intelligence cs ai machine learning cs lg arxiv link pdf link abstract recently it has been exposed that some modern facial recognition systems could discriminate specific demographic groups and may lead to unfair attention with respect to various facial attributes such as gender and origin the main reason are the biases inside datasets unbalanced demographics used to train theses models unfortunately collecting a large scale balanced dataset with respect to various demographics is impracticable in this paper we investigate as an alternative the generation of a balanced and possibly bias free synthetic dataset that could be used to train to regularize or to evaluate deep learning based facial recognition models we propose to use a simple method for modeling and sampling a disentangled projection of a stylegan latent space to generate any combination of demographic groups e g hispanic female our experiments show that we can synthesis any combination of demographic groups effectively and the identities are different from the original training dataset we also released the source code visual speech recognition for low resource languages with automatic labels from whisper model authors jeong hun yeo minsu kim shinji watanabe yong man ro subjects computer vision and pattern recognition cs cv artificial intelligence cs ai audio and speech processing eess as arxiv link pdf link abstract this paper proposes a powerful visual speech recognition vsr method for multiple languages especially for low resource languages that have a limited number of labeled data different from previous methods that tried to improve the vsr performance for the target language by using knowledge learned from other languages we explore whether we can increase the amount of training data itself for the different languages without human intervention to this end we employ a whisper model which can conduct both language identification and audio based speech recognition it serves to filter data of the desired languages and transcribe labels from the unannotated multilingual audio visual data pool by comparing the performances of vsr models trained on automatic labels and the human annotated labels we show that we can achieve similar vsr performance to that of human annotated labels even without utilizing human annotations through the automated labeling process we label large scale unlabeled multilingual databases and avspeech producing hours of data for four low vsr resource languages french italian spanish and portuguese with the automatic labels we achieve new state of the art performance on mtedx in four languages significantly surpassing the previous methods the automatic labels are available online keyword image signal processing let s roll synthetic dataset analysis for pedestrian detection across different shutter types authors yue hu gourav datta kira beerel peter beerel subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract computer vision cv pipelines are typically evaluated on datasets processed by image signal processing isp pipelines even though for resource constrained applications an important research goal is to avoid as many isp steps as possible in particular most cv datasets consist of global shutter gs images even though most cameras today use a rolling shutter rs this paper studies the impact of different shutter mechanisms on machine learning ml object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of unreal engine in particular we train and evaluate mainstream detection models with our synthetically generated paired gs and rs datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities especially when capturing low speed objects e g pedestrians the results of this emulation framework indicate the performance between them are remarkably congruent for coarse grained detection mean average precision map for iou but have significant differences for fine grained measures of detection accuracy map for iou this implies that ml pipelines might not need explicit correction for rs for many object detection applications but mitigating rs effects in isp less ml pipelines that target fine grained location of the objects may need additional research keyword image signal process let s roll synthetic dataset analysis for pedestrian detection across different shutter types authors yue hu gourav datta kira beerel peter beerel subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract computer vision cv pipelines are typically evaluated on datasets processed by image signal processing isp pipelines even though for resource constrained applications an important research goal is to avoid as many isp steps as possible in particular most cv datasets consist of global shutter gs images even though most cameras today use a rolling shutter rs this paper studies the impact of different shutter mechanisms on machine learning ml object detection models on a synthetic dataset that we generate using the advanced simulation capabilities of unreal engine in particular we train and evaluate mainstream detection models with our synthetically generated paired gs and rs datasets to ascertain whether there exists a significant difference in detection accuracy between these two shutter modalities especially when capturing low speed objects e g pedestrians the results of this emulation framework indicate the performance between them are remarkably congruent for coarse grained detection mean average precision map for iou but have significant differences for fine grained measures of detection accuracy map for iou this implies that ml pipelines might not need explicit correction for rs for many object detection applications but mitigating rs effects in isp less ml pipelines that target fine grained location of the objects may need additional research keyword compression differentiable resolution compression and alignment for efficient video classification and retrieval authors rui deng qian wu yuke li haoran fu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract optimizing video inference efficiency has become increasingly important with the growing demand for video analysis in various fields some existing methods achieve high efficiency by explicit discard of spatial or temporal information which poses challenges in fast changing and fine grained scenarios to address these issues we propose an efficient video representation network with differentiable resolution compression and alignment mechanism which compresses non essential information in the early stage of the network to reduce computational costs while maintaining consistent temporal correlations specifically we leverage a differentiable context aware compression module to encode the saliency and non saliency frame features refining and updating the features into a high low resolution video sequence to process the new sequence we introduce a new resolution align transformer layer to capture global temporal correlations among frame features with different resolutions while reducing spatial computation costs quadratically by utilizing fewer spatial tokens in low resolution non saliency frames the entire network can be end to end optimized via the integration of the differentiable compression module experimental results show that our method achieves the best trade off between efficiency and performance on near duplicate video retrieval and competitive results on dynamic video classification compared to state of the art methods code keyword raw interpretability aware vision transformer authors yao qiang chengyin li prashant khanduri dongxiao zhu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract vision transformers vits have become prominent models for solving various vision tasks however the interpretability of vits has not kept pace with their promising performance while there has been a surge of interest in developing it post hoc solutions to explain vits outputs these methods do not generalize to different downstream tasks and various transformer architectures furthermore if vits are not properly trained with the given data and do not prioritize the region of interest the it post hoc methods would be less effective instead of developing another it post hoc approach we introduce a novel training procedure that inherently enhances model interpretability our interpretability aware vit ia vit draws inspiration from a fresh insight both the class patch and image patches consistently generate predicted distributions and attention maps ia vit is composed of a feature extractor a predictor and an interpreter which are trained jointly with an interpretability aware training objective consequently the interpreter simulates the behavior of the predictor and provides a faithful explanation through its single head self attention mechanism our comprehensive experimental results demonstrate the effectiveness of ia vit in several image classification tasks with both qualitative and quantitative evaluations of model performance and interpretability source code is available from salient object detection in optical remote sensing images driven by transformer authors gongyang li zhen bai zhi liu xinpeng zhang haibin ling subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract existing methods for salient object detection in optical remote sensing images orsi sod mainly adopt convolutional neural networks cnns as the backbone such as vgg and resnet since cnns can only extract features within certain receptive fields most orsi sod methods generally follow the local to contextual paradigm in this paper we propose a novel global extraction local exploration network gelenet for orsi sod following the global to local paradigm specifically gelenet first adopts a transformer backbone to generate four level feature embeddings with global long range dependencies then gelenet employs a direction aware shuffle weighted spatial attention module d swsam and its simplified version swsam to enhance local interactions and a knowledge transfer module ktm to further enhance cross level contextual interactions d swsam comprehensively perceives the orientation information in the lowest level features through directional convolutions to adapt to various orientations of salient objects in orsis and effectively enhances the details of salient objects with an improved attention mechanism swsam discards the direction aware part of d swsam to focus on localizing salient objects in the highest level features ktm models the contextual correlation knowledge of two middle level features of different scales based on the self attention mechanism and transfers the knowledge to the raw features to generate more discriminative features finally a saliency predictor is used to generate the saliency map based on the outputs of the above three modules extensive experiments on three public datasets demonstrate that the proposed gelenet outperforms relevant state of the art methods the code and results of our method are available at beyond domain gap exploiting subjectivity in sketch based person retrieval authors kejun lin zhixiang wang zheng wang yinqiang zheng shin ichi satoh subjects computer vision and pattern recognition cs cv multimedia cs mm arxiv link pdf link abstract person re identification re id requires densely distributed cameras in practice the person of interest may not be captured by cameras and therefore needs to be retrieved using subjective information e g sketches from witnesses previous research defines this case using the sketch as sketch re identification sketch re id and focuses on eliminating the domain gap actually subjectivity is another significant challenge we model and investigate it by posing a new dataset with multi witness descriptions it features two aspects large scale it contains over sketches and photos making it the largest sketch re id dataset multi perspective and multi style our dataset offers multiple sketches for each identity witnesses subjective cognition provides multiple perspectives on the same individual while different artists drawing styles provide variation in sketch styles we further have two novel designs to alleviate the challenge of subjectivity fusing subjectivity we propose a non local nl fusion module that gathers sketches from different witnesses for the same identity introducing objectivity an attralign module utilizes attributes as an implicit mask to align cross domain features to push forward the advance of sketch re id we set three benchmarks large scale multi style cross style extensive experiments demonstrate our leading performance in these benchmarks dataset and codes are publicly available at towards practical and efficient image to speech captioning with vision language pre training and multi modal tokens authors minsu kim jeongsoo choi soumi maiti jeong hun yeo shinji watanabe yong man ro subjects computer vision and pattern recognition cs cv computation and language cs cl audio and speech processing eess as image and video processing eess iv arxiv link pdf link abstract in this paper we propose methods to build a powerful and efficient image to speech captioning model to this end we start with importing the rich knowledge related to image comprehension and language modeling from a large scale pre trained vision language model into we set the output of the proposed as discretized speech units i e the quantized speech features of a self supervised speech model the speech units mainly contain linguistic information while suppressing other characteristics of speech this allows us to incorporate the language modeling capability of the pre trained vision language model into the spoken language modeling of with the vision language pre training strategy we set new state of the art performances on two widely used benchmark databases coco and then we further improve the efficiency of the model similar to the speech unit case we convert the original image into image units which are derived through vector quantization of the raw image with these image units we can drastically reduce the required data storage for saving image data to just when compared to the original image data in terms of bits demo page keyword raw image towards practical and efficient image to speech captioning with vision language pre training and multi modal tokens authors minsu kim jeongsoo choi soumi maiti jeong hun yeo shinji watanabe yong man ro subjects computer vision and pattern recognition cs cv computation and language cs cl audio and speech processing eess as image and video processing eess iv arxiv link pdf link abstract in this paper we propose methods to build a powerful and efficient image to speech captioning model to this end we start with importing the rich knowledge related to image comprehension and language modeling from a large scale pre trained vision language model into we set the output of the proposed as discretized speech units i e the quantized speech features of a self supervised speech model the speech units mainly contain linguistic information while suppressing other characteristics of speech this allows us to incorporate the language modeling capability of the pre trained vision language model into the spoken language modeling of with the vision language pre training strategy we set new state of the art performances on two widely used benchmark databases coco and then we further improve the efficiency of the model similar to the speech unit case we convert the original image into image units which are derived through vector quantization of the raw image with these image units we can drastically reduce the required data storage for saving image data to just when compared to the original image data in terms of bits demo page
1
153,751
12,159,976,090
IssuesEvent
2020-04-26 11:25:35
AdventistCommons/adventistcommons.org
https://api.github.com/repos/AdventistCommons/adventistcommons.org
closed
RTL Languages
To test question
Do you think we should make all RTL languages unavailable for now because we don't support them right now?
1.0
RTL Languages - Do you think we should make all RTL languages unavailable for now because we don't support them right now?
non_process
rtl languages do you think we should make all rtl languages unavailable for now because we don t support them right now
0
12,919
15,294,219,834
IssuesEvent
2021-02-24 01:59:56
carterclark/GroceryStore
https://api.github.com/repos/carterclark/GroceryStore
closed
Enroll a member
Business Process
The system enrolls a member and remembers him/her; it keeps track of the name, address, phone number, date joined, and the fee paid. Also, the system creates and maintains a unique id for each member. Only one member is added when this functionality is invoked.
1.0
Enroll a member - The system enrolls a member and remembers him/her; it keeps track of the name, address, phone number, date joined, and the fee paid. Also, the system creates and maintains a unique id for each member. Only one member is added when this functionality is invoked.
process
enroll a member the system enrolls a member and remembers him her it keeps track of the name address phone number date joined and the fee paid also the system creates and maintains a unique id for each member only one member is added when this functionality is invoked
1
93
2,589,940,883
IssuesEvent
2015-02-18 16:02:18
betterFORM/ziziphus
https://api.github.com/repos/betterFORM/ziziphus
closed
[agents]: VIAF integration
feature request important services vocabularies
For all agents, the main repository to reference should be VIAF (cf http://viaf.org). To allow the offline use of VIAF, a "local" instance should be implemented. In addition, the online VIAF database should be made accessible for search from within Ziziphus. The suggestion is to open the VIAF interface within an i-frame, where the user can search and identifiy the correct agent. The ID of the agent selected by the user should be referenced in the VRA record.
1.0
[agents]: VIAF integration - For all agents, the main repository to reference should be VIAF (cf http://viaf.org). To allow the offline use of VIAF, a "local" instance should be implemented. In addition, the online VIAF database should be made accessible for search from within Ziziphus. The suggestion is to open the VIAF interface within an i-frame, where the user can search and identifiy the correct agent. The ID of the agent selected by the user should be referenced in the VRA record.
non_process
viaf integration for all agents the main repository to reference should be viaf cf to allow the offline use of viaf a local instance should be implemented in addition the online viaf database should be made accessible for search from within ziziphus the suggestion is to open the viaf interface within an i frame where the user can search and identifiy the correct agent the id of the agent selected by the user should be referenced in the vra record
0
18,520
24,551,784,411
IssuesEvent
2022-10-12 13:09:11
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [Offline indicator] Activities screen is continuously loading in the following scenario
Bug P0 iOS Process: Fixed Process: Tested dev
Steps: 1. Sign up or sign in to the mobile app 2. Enroll to the study 3. In SB, update the consent for enrolled users 4. In the mobile app, Click on the enrolled study 5. Turn OFF the internet 6. Click on Review and turn ON the internet 7. Click on refresh the page 8. Repeat the step 5 and step 6 9. Observe AR: Activities screen is loading continuously ER: Activities should get loaded and displayed to the participants ![image](https://user-images.githubusercontent.com/71445210/185437136-8c0f0c51-3834-4a04-b354-e19370e1908a.png)
2.0
[iOS] [Offline indicator] Activities screen is continuously loading in the following scenario - Steps: 1. Sign up or sign in to the mobile app 2. Enroll to the study 3. In SB, update the consent for enrolled users 4. In the mobile app, Click on the enrolled study 5. Turn OFF the internet 6. Click on Review and turn ON the internet 7. Click on refresh the page 8. Repeat the step 5 and step 6 9. Observe AR: Activities screen is loading continuously ER: Activities should get loaded and displayed to the participants ![image](https://user-images.githubusercontent.com/71445210/185437136-8c0f0c51-3834-4a04-b354-e19370e1908a.png)
process
activities screen is continuously loading in the following scenario steps sign up or sign in to the mobile app enroll to the study in sb update the consent for enrolled users in the mobile app click on the enrolled study turn off the internet click on review and turn on the internet click on refresh the page repeat the step and step observe ar activities screen is loading continuously er activities should get loaded and displayed to the participants
1
166,321
14,047,513,535
IssuesEvent
2020-11-02 07:18:22
thephpleague/uri
https://api.github.com/repos/thephpleague/uri
closed
League\Uri\Components\Host is missing
documentation question
### Bug Report On this page is explained how you can analyze a host: https://github.com/thephpleague/uri/blob/master/docs/index.md However, the class `League\Uri\Components\Host` does not exists in the latest version. Is this functionality completely lost in version 6?
1.0
League\Uri\Components\Host is missing - ### Bug Report On this page is explained how you can analyze a host: https://github.com/thephpleague/uri/blob/master/docs/index.md However, the class `League\Uri\Components\Host` does not exists in the latest version. Is this functionality completely lost in version 6?
non_process
league uri components host is missing bug report on this page is explained how you can analyze a host however the class league uri components host does not exists in the latest version is this functionality completely lost in version
0
6,059
8,881,852,304
IssuesEvent
2019-01-14 11:23:20
w3c/payment-request
https://api.github.com/repos/w3c/payment-request
closed
Confirm tests and specs align
Editorial Process aid
We need to make sure that the tests listed in the spec and the actual tests in the test suite align. We need: 1. The list of files in WPT's [payment request](https://github.com/web-platform-tests/wpt/tree/master/payment-request) directory. 1. To be compared to the "data-test" files listed in [the spec](https://w3c.github.io/payment-request/) ([example](https://github.com/w3c/payment-request/blob/gh-pages/index.html#L4898)). Then we can figure out: * Are all tests in the test suite in the spec? * Are tests listed in the spec that don't actually link to anything? @Sylvia23 has offered to help us with this 👋 The tests that are in the payment request directory but not in the spec are: - [x] MerchantValidationEvent/complete-method.https.html - [ ] PaymentAddress/attributes-and-toJSON-method-manual.https.html - [ ] PaymentMethodChangeEvent/methodDetails-attribute.https.html - [ ] PaymentMethodChangeEvent/methodName-attribute.https.html - [ ] PaymentRequestUpdateEvent/updateWith-call-immediate-manual.https.html - [ ] PaymentRequestUpdateEvent/updateWith-method-abort-update-manual.https.html - [ ] PaymentRequestUpdateEvent/updateWith-state-checks-manual.https.html - [ ] PaymentRequestUpdateEvent/updatewith-method.https.html - [ ] PaymentValidationErrors/retry-shows-payer-member-manual.https.html - [ ] PaymentValidationErrors/retry-shows-shippingAddress-member-manual.https.html - [ ] payment-response/onpayerdetailchange-attribute-manual.https.html - [ ] payment-response/onpayerdetailchange-attribute.https.html - [x] payment-request-abort-method.https.html - [x] payment-request-canmakepayment-method.https.html - [ ] payment-request-not-exposed.https.worker.js -The tests that are in the spec but are not present in or are wrongly addressed in the payment directory are: ### Line 640: - [x] active-document-cross-origin.https.sub.html - [x] active-document-same-origin.https.html - [x] removing-allowpaymentrequest.https.sub.html - [x] setting-allowpaymentrequest-timing.https.sub.html - [x] setting-allowpaymentrequest.https.sub.html ### Line 904: - [x] payment-request-show-method-manual.https.html ### Line 910: - [x] payment-request-show-method-manual.https.html ### Line 1138: - [x] payment-request-abort-method-manual.https.html ### Line 1197: - [x] payment-request-canmakepayment-method-manual.https.html ### Line 1300: - [x] payment-request/onmerchantvalidation-attribute.https.html ### Line 3290: - [x] payment-request/PaymentValidationErrors/retry-shows-error-member-manual.https.html ### Line 3322: - [x] payment-request/payment-response/rejects_if_not_active-manual.https.html ### Line 3630: - [x] payment-request/payment-response/retry-method-manual.https.html ### Line 4177: - [x] payment-request-update-event-updatewith-method.https.html
1.0
Confirm tests and specs align - We need to make sure that the tests listed in the spec and the actual tests in the test suite align. We need: 1. The list of files in WPT's [payment request](https://github.com/web-platform-tests/wpt/tree/master/payment-request) directory. 1. To be compared to the "data-test" files listed in [the spec](https://w3c.github.io/payment-request/) ([example](https://github.com/w3c/payment-request/blob/gh-pages/index.html#L4898)). Then we can figure out: * Are all tests in the test suite in the spec? * Are tests listed in the spec that don't actually link to anything? @Sylvia23 has offered to help us with this 👋 The tests that are in the payment request directory but not in the spec are: - [x] MerchantValidationEvent/complete-method.https.html - [ ] PaymentAddress/attributes-and-toJSON-method-manual.https.html - [ ] PaymentMethodChangeEvent/methodDetails-attribute.https.html - [ ] PaymentMethodChangeEvent/methodName-attribute.https.html - [ ] PaymentRequestUpdateEvent/updateWith-call-immediate-manual.https.html - [ ] PaymentRequestUpdateEvent/updateWith-method-abort-update-manual.https.html - [ ] PaymentRequestUpdateEvent/updateWith-state-checks-manual.https.html - [ ] PaymentRequestUpdateEvent/updatewith-method.https.html - [ ] PaymentValidationErrors/retry-shows-payer-member-manual.https.html - [ ] PaymentValidationErrors/retry-shows-shippingAddress-member-manual.https.html - [ ] payment-response/onpayerdetailchange-attribute-manual.https.html - [ ] payment-response/onpayerdetailchange-attribute.https.html - [x] payment-request-abort-method.https.html - [x] payment-request-canmakepayment-method.https.html - [ ] payment-request-not-exposed.https.worker.js -The tests that are in the spec but are not present in or are wrongly addressed in the payment directory are: ### Line 640: - [x] active-document-cross-origin.https.sub.html - [x] active-document-same-origin.https.html - [x] removing-allowpaymentrequest.https.sub.html - [x] setting-allowpaymentrequest-timing.https.sub.html - [x] setting-allowpaymentrequest.https.sub.html ### Line 904: - [x] payment-request-show-method-manual.https.html ### Line 910: - [x] payment-request-show-method-manual.https.html ### Line 1138: - [x] payment-request-abort-method-manual.https.html ### Line 1197: - [x] payment-request-canmakepayment-method-manual.https.html ### Line 1300: - [x] payment-request/onmerchantvalidation-attribute.https.html ### Line 3290: - [x] payment-request/PaymentValidationErrors/retry-shows-error-member-manual.https.html ### Line 3322: - [x] payment-request/payment-response/rejects_if_not_active-manual.https.html ### Line 3630: - [x] payment-request/payment-response/retry-method-manual.https.html ### Line 4177: - [x] payment-request-update-event-updatewith-method.https.html
process
confirm tests and specs align we need to make sure that the tests listed in the spec and the actual tests in the test suite align we need the list of files in wpt s directory to be compared to the data test files listed in then we can figure out are all tests in the test suite in the spec are tests listed in the spec that don t actually link to anything has offered to help us with this 👋 the tests that are in the payment request directory but not in the spec are merchantvalidationevent complete method https html paymentaddress attributes and tojson method manual https html paymentmethodchangeevent methoddetails attribute https html paymentmethodchangeevent methodname attribute https html paymentrequestupdateevent updatewith call immediate manual https html paymentrequestupdateevent updatewith method abort update manual https html paymentrequestupdateevent updatewith state checks manual https html paymentrequestupdateevent updatewith method https html paymentvalidationerrors retry shows payer member manual https html paymentvalidationerrors retry shows shippingaddress member manual https html payment response onpayerdetailchange attribute manual https html payment response onpayerdetailchange attribute https html payment request abort method https html payment request canmakepayment method https html payment request not exposed https worker js the tests that are in the spec but are not present in or are wrongly addressed in the payment directory are line active document cross origin https sub html active document same origin https html removing allowpaymentrequest https sub html setting allowpaymentrequest timing https sub html setting allowpaymentrequest https sub html line payment request show method manual https html line payment request show method manual https html line payment request abort method manual https html line payment request canmakepayment method manual https html line payment request onmerchantvalidation attribute https html line payment request paymentvalidationerrors retry shows error member manual https html line payment request payment response rejects if not active manual https html line payment request payment response retry method manual https html line payment request update event updatewith method https html
1
22,080
30,603,420,666
IssuesEvent
2023-07-22 17:34:24
rust-lang/cargo
https://api.github.com/repos/rust-lang/cargo
closed
Error message with `credential-process` and a `token` is confusing.
C-bug A-errors S-needs-mentor Z-credential-process
### Problem If you are currently logged in to a registry using the typical `token` setting in `credentials.toml`, and you add a configuration for `registry.credential-process`, the error message when running `cargo login` is a little confusing: ``` error: both `token` and `credential-process` were specified in the config. Only one of these values may be set, remove one or the other to proceed. ``` If a user is not following the instructions to the letter, or is not intimately familiar with cargo's auth setup, this doesn't really tell you what to do to fix the error. ### Steps 1. Do a normal login: `cargo login` 2. Set up a credential process in `config.toml`: ```toml [registry] credential-process = "cargo:1password" ``` 3. Run `cargo login -Zcredential-process` ### Possible Solution(s) The error message could be a little clearer about what it means about a "`token`...in the config". A user may not know or equate that a legacy `cargo login` means that a "token is in the config". I would probably completely reword this error message to explain that they are logged in via the old mechanism, and that needs to be removed before using `credential-process` (though I'm not sure how to phrase "old mechanism"). I think the message could also include some instructions on how to fix it (such as running `cargo logout`). ### Notes _No response_ ### Version ```text cargo 1.72.0-nightly (03bc66b55 2023-06-23) release: 1.72.0-nightly commit-hash: 03bc66b55c290324bd46eb22e369c8fae1908f91 commit-date: 2023-06-23 host: aarch64-unknown-linux-gnu libgit2: 1.6.4 (sys:0.17.2 vendored) libcurl: 8.1.2-DEV (sys:0.4.63+curl-8.1.2 vendored ssl:OpenSSL/1.1.1u) ssl: OpenSSL 1.1.1u 30 May 2023 os: Ubuntu 22.04 (jammy) [64-bit] ``` <!-- TRIAGEBOT_START --> <!-- TRIAGEBOT_ASSIGN_START --> <!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"chansuke"}$$TRIAGEBOT_ASSIGN_DATA_END --> <!-- TRIAGEBOT_ASSIGN_END --> <!-- TRIAGEBOT_END -->
1.0
Error message with `credential-process` and a `token` is confusing. - ### Problem If you are currently logged in to a registry using the typical `token` setting in `credentials.toml`, and you add a configuration for `registry.credential-process`, the error message when running `cargo login` is a little confusing: ``` error: both `token` and `credential-process` were specified in the config. Only one of these values may be set, remove one or the other to proceed. ``` If a user is not following the instructions to the letter, or is not intimately familiar with cargo's auth setup, this doesn't really tell you what to do to fix the error. ### Steps 1. Do a normal login: `cargo login` 2. Set up a credential process in `config.toml`: ```toml [registry] credential-process = "cargo:1password" ``` 3. Run `cargo login -Zcredential-process` ### Possible Solution(s) The error message could be a little clearer about what it means about a "`token`...in the config". A user may not know or equate that a legacy `cargo login` means that a "token is in the config". I would probably completely reword this error message to explain that they are logged in via the old mechanism, and that needs to be removed before using `credential-process` (though I'm not sure how to phrase "old mechanism"). I think the message could also include some instructions on how to fix it (such as running `cargo logout`). ### Notes _No response_ ### Version ```text cargo 1.72.0-nightly (03bc66b55 2023-06-23) release: 1.72.0-nightly commit-hash: 03bc66b55c290324bd46eb22e369c8fae1908f91 commit-date: 2023-06-23 host: aarch64-unknown-linux-gnu libgit2: 1.6.4 (sys:0.17.2 vendored) libcurl: 8.1.2-DEV (sys:0.4.63+curl-8.1.2 vendored ssl:OpenSSL/1.1.1u) ssl: OpenSSL 1.1.1u 30 May 2023 os: Ubuntu 22.04 (jammy) [64-bit] ``` <!-- TRIAGEBOT_START --> <!-- TRIAGEBOT_ASSIGN_START --> <!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"chansuke"}$$TRIAGEBOT_ASSIGN_DATA_END --> <!-- TRIAGEBOT_ASSIGN_END --> <!-- TRIAGEBOT_END -->
process
error message with credential process and a token is confusing problem if you are currently logged in to a registry using the typical token setting in credentials toml and you add a configuration for registry credential process the error message when running cargo login is a little confusing error both token and credential process were specified in the config only one of these values may be set remove one or the other to proceed if a user is not following the instructions to the letter or is not intimately familiar with cargo s auth setup this doesn t really tell you what to do to fix the error steps do a normal login cargo login set up a credential process in config toml toml credential process cargo run cargo login zcredential process possible solution s the error message could be a little clearer about what it means about a token in the config a user may not know or equate that a legacy cargo login means that a token is in the config i would probably completely reword this error message to explain that they are logged in via the old mechanism and that needs to be removed before using credential process though i m not sure how to phrase old mechanism i think the message could also include some instructions on how to fix it such as running cargo logout notes no response version text cargo nightly release nightly commit hash commit date host unknown linux gnu sys vendored libcurl dev sys curl vendored ssl openssl ssl openssl may os ubuntu jammy
1
1,421
3,987,666,368
IssuesEvent
2016-05-09 05:47:27
kerubistan/kerub
https://api.github.com/repos/kerubistan/kerub
closed
exception when uploading files > 2GB
bug component:data processing
May be a timeout. ``` Caused by: org.apache.sshd.common.subsystem.sftp.SftpException: Failure at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.throwStatusException(AbstractSftpClient.java:254) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.checkResponseStatus(AbstractSftpClient.java:249) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.checkResponseStatus(AbstractSftpClient.java:226) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.checkCommandStatus(AbstractSftpClient.java:205) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.write(AbstractSftpClient.java:904) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.SftpOutputStreamWithChannel.flush(SftpOutputStreamWithChannel.java:103) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.SftpOutputStreamWithChannel.close(SftpOutputStreamWithChannel.java:114) ~[sshd-core-1.2.0.jar:1.2.0] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$pump$1.invoke(VirtualStorageDeviceServiceImpl.kt:88) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$pump$1.invoke(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.host.SshClientUtilsKt.use(SshClientUtils.kt:114) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl.pump(VirtualStorageDeviceServiceImpl.kt:52) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl.access$pump(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1$1.invoke(VirtualStorageDeviceServiceImpl.kt:35) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1$1.invoke(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.host.HostManagerImpl.dataConnection(HostManagerImpl.kt:90) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1.invoke(VirtualStorageDeviceServiceImpl.kt:33) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1.invoke(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$waitFor$1.invoke(IspnDaoBase.kt:132) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$waitFor$1.invoke(IspnDaoBase.kt:23) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$AbstractListener.onEvent(IspnDaoBase.kt:52) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$CreateListener.onCreate(IspnDaoBase.kt:61) ~[classes/:na] ... 8 common frames omitted ```
1.0
exception when uploading files > 2GB - May be a timeout. ``` Caused by: org.apache.sshd.common.subsystem.sftp.SftpException: Failure at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.throwStatusException(AbstractSftpClient.java:254) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.checkResponseStatus(AbstractSftpClient.java:249) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.checkResponseStatus(AbstractSftpClient.java:226) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.checkCommandStatus(AbstractSftpClient.java:205) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.AbstractSftpClient.write(AbstractSftpClient.java:904) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.SftpOutputStreamWithChannel.flush(SftpOutputStreamWithChannel.java:103) ~[sshd-core-1.2.0.jar:1.2.0] at org.apache.sshd.client.subsystem.sftp.SftpOutputStreamWithChannel.close(SftpOutputStreamWithChannel.java:114) ~[sshd-core-1.2.0.jar:1.2.0] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$pump$1.invoke(VirtualStorageDeviceServiceImpl.kt:88) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$pump$1.invoke(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.host.SshClientUtilsKt.use(SshClientUtils.kt:114) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl.pump(VirtualStorageDeviceServiceImpl.kt:52) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl.access$pump(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1$1.invoke(VirtualStorageDeviceServiceImpl.kt:35) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1$1.invoke(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.host.HostManagerImpl.dataConnection(HostManagerImpl.kt:90) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1.invoke(VirtualStorageDeviceServiceImpl.kt:33) ~[classes/:na] at com.github.K0zka.kerub.services.impl.VirtualStorageDeviceServiceImpl$load$1.invoke(VirtualStorageDeviceServiceImpl.kt:20) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$waitFor$1.invoke(IspnDaoBase.kt:132) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$waitFor$1.invoke(IspnDaoBase.kt:23) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$AbstractListener.onEvent(IspnDaoBase.kt:52) ~[classes/:na] at com.github.K0zka.kerub.data.ispn.IspnDaoBase$CreateListener.onCreate(IspnDaoBase.kt:61) ~[classes/:na] ... 8 common frames omitted ```
process
exception when uploading files may be a timeout caused by org apache sshd common subsystem sftp sftpexception failure at org apache sshd client subsystem sftp abstractsftpclient throwstatusexception abstractsftpclient java at org apache sshd client subsystem sftp abstractsftpclient checkresponsestatus abstractsftpclient java at org apache sshd client subsystem sftp abstractsftpclient checkresponsestatus abstractsftpclient java at org apache sshd client subsystem sftp abstractsftpclient checkcommandstatus abstractsftpclient java at org apache sshd client subsystem sftp abstractsftpclient write abstractsftpclient java at org apache sshd client subsystem sftp sftpoutputstreamwithchannel flush sftpoutputstreamwithchannel java at org apache sshd client subsystem sftp sftpoutputstreamwithchannel close sftpoutputstreamwithchannel java at com github kerub services impl virtualstoragedeviceserviceimpl pump invoke virtualstoragedeviceserviceimpl kt at com github kerub services impl virtualstoragedeviceserviceimpl pump invoke virtualstoragedeviceserviceimpl kt at com github kerub host sshclientutilskt use sshclientutils kt at com github kerub services impl virtualstoragedeviceserviceimpl pump virtualstoragedeviceserviceimpl kt at com github kerub services impl virtualstoragedeviceserviceimpl access pump virtualstoragedeviceserviceimpl kt at com github kerub services impl virtualstoragedeviceserviceimpl load invoke virtualstoragedeviceserviceimpl kt at com github kerub services impl virtualstoragedeviceserviceimpl load invoke virtualstoragedeviceserviceimpl kt at com github kerub host hostmanagerimpl dataconnection hostmanagerimpl kt at com github kerub services impl virtualstoragedeviceserviceimpl load invoke virtualstoragedeviceserviceimpl kt at com github kerub services impl virtualstoragedeviceserviceimpl load invoke virtualstoragedeviceserviceimpl kt at com github kerub data ispn ispndaobase waitfor invoke ispndaobase kt at com github kerub data ispn ispndaobase waitfor invoke ispndaobase kt at com github kerub data ispn ispndaobase abstractlistener onevent ispndaobase kt at com github kerub data ispn ispndaobase createlistener oncreate ispndaobase kt common frames omitted
1
74,549
14,269,212,641
IssuesEvent
2020-11-21 00:34:34
openshift/odo
https://api.github.com/repos/openshift/odo
closed
Switch to use the parser from devfile/parser in odo repo
area/devfile kind/code-refactoring
After https://github.com/devfile/api/issues/183 is completed, we want to switch to use the parser from devfile/parser repo, and remove the parser defined in Odo. Since the parser library is using devfile go struct defined in devfile/api, this requires some code changes to Odo code base to integrate with the new devfile go struct. product code and test code changes are expected. **Note: the validate package will temporarily be left in odo, the validator will be implemented in devfile/api repo. For any call to the following parser functions in odo, need to call `validate.ValidateDevfileData` after, separately: `ParseAndValidate`, `ParseFromURLAndValidate`, `ParseFromDataAndValidate`,
1.0
Switch to use the parser from devfile/parser in odo repo - After https://github.com/devfile/api/issues/183 is completed, we want to switch to use the parser from devfile/parser repo, and remove the parser defined in Odo. Since the parser library is using devfile go struct defined in devfile/api, this requires some code changes to Odo code base to integrate with the new devfile go struct. product code and test code changes are expected. **Note: the validate package will temporarily be left in odo, the validator will be implemented in devfile/api repo. For any call to the following parser functions in odo, need to call `validate.ValidateDevfileData` after, separately: `ParseAndValidate`, `ParseFromURLAndValidate`, `ParseFromDataAndValidate`,
non_process
switch to use the parser from devfile parser in odo repo after is completed we want to switch to use the parser from devfile parser repo and remove the parser defined in odo since the parser library is using devfile go struct defined in devfile api this requires some code changes to odo code base to integrate with the new devfile go struct product code and test code changes are expected note the validate package will temporarily be left in odo the validator will be implemented in devfile api repo for any call to the following parser functions in odo need to call validate validatedevfiledata after separately parseandvalidate parsefromurlandvalidate parsefromdataandvalidate
0
138,608
12,824,782,329
IssuesEvent
2020-07-06 14:00:33
double-great/alt-text
https://api.github.com/repos/double-great/alt-text
closed
Include sounds files in documentation
documentation
We should consider adding sounds files of a screen reader reading alt text in the documentation. This can help the developer hear the differences between the recommended and discouraged alt text.
1.0
Include sounds files in documentation - We should consider adding sounds files of a screen reader reading alt text in the documentation. This can help the developer hear the differences between the recommended and discouraged alt text.
non_process
include sounds files in documentation we should consider adding sounds files of a screen reader reading alt text in the documentation this can help the developer hear the differences between the recommended and discouraged alt text
0
1,677
4,313,249,485
IssuesEvent
2016-07-22 09:40:55
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Customize a new connection with session variables
ADMIN CONNECTION POOL MYSQL PROTOCOL QUERY PROCESSOR
When ProxySQL connects to a backend it should be able to define custom variables, like `wait_timeout`, `sql_mode`, etc
1.0
Customize a new connection with session variables - When ProxySQL connects to a backend it should be able to define custom variables, like `wait_timeout`, `sql_mode`, etc
process
customize a new connection with session variables when proxysql connects to a backend it should be able to define custom variables like wait timeout sql mode etc
1
17,384
12,335,964,799
IssuesEvent
2020-05-14 12:52:58
google/web-stories-wp
https://api.github.com/repos/google/web-stories-wp
closed
Local build sometimes fails when running in watch mode
Status: Stale Type: Infrastructure
## Bug Description When running `npm run dev` local build fails spontaneously. However, it does not fail in the terminal, the error is only reported as a failed build when reloading the browser window. The failed builds occur regardless of the change. Simply change any line anywhere can cause the build to fail - changing a `1` to a `2` in any file back and forth will trigger the error after a few attempts. Aborting the watching process and restarting it always produces a valid build. ## Expected Behaviour The build should never fail in the browser only. ## Steps to Reproduce 1. Checkout `master` 1. Run `npm i` 1. Run `npm run dev` 1. Make changes to any file and watch process rebuild 1. Reload browser to see resulting changes 1. The browser renders nothing and throws an error: ``` Uncaught Error: Module build failed (from ./node_modules/babel-loader/lib/index.js): Error: <root>/assets/src/edit-story/components/canvas/gridview/index.js: ENOENT: no such file or directory, open '<root>/node_modules/svgo/lib/svgo/../../.svgo.yml' ``` The actual file containing the error (in the above `components/canvas/gridview/index.js`) can vary - it can be any file containing a reference to an SVG. And the file is not related to the change performed triggering the rebuild. 7. If no error happened, repeat from step 4 until one does occur. It takes no more than 4 attempts on my setup. ## Screenshots <!-- If applicable, please add screenshots to help explain your problem. Bonus points for videos! --> --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria ## QA Instructions
1.0
Local build sometimes fails when running in watch mode - ## Bug Description When running `npm run dev` local build fails spontaneously. However, it does not fail in the terminal, the error is only reported as a failed build when reloading the browser window. The failed builds occur regardless of the change. Simply change any line anywhere can cause the build to fail - changing a `1` to a `2` in any file back and forth will trigger the error after a few attempts. Aborting the watching process and restarting it always produces a valid build. ## Expected Behaviour The build should never fail in the browser only. ## Steps to Reproduce 1. Checkout `master` 1. Run `npm i` 1. Run `npm run dev` 1. Make changes to any file and watch process rebuild 1. Reload browser to see resulting changes 1. The browser renders nothing and throws an error: ``` Uncaught Error: Module build failed (from ./node_modules/babel-loader/lib/index.js): Error: <root>/assets/src/edit-story/components/canvas/gridview/index.js: ENOENT: no such file or directory, open '<root>/node_modules/svgo/lib/svgo/../../.svgo.yml' ``` The actual file containing the error (in the above `components/canvas/gridview/index.js`) can vary - it can be any file containing a reference to an SVG. And the file is not related to the change performed triggering the rebuild. 7. If no error happened, repeat from step 4 until one does occur. It takes no more than 4 attempts on my setup. ## Screenshots <!-- If applicable, please add screenshots to help explain your problem. Bonus points for videos! --> --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance Criteria ## QA Instructions
non_process
local build sometimes fails when running in watch mode bug description when running npm run dev local build fails spontaneously however it does not fail in the terminal the error is only reported as a failed build when reloading the browser window the failed builds occur regardless of the change simply change any line anywhere can cause the build to fail changing a to a in any file back and forth will trigger the error after a few attempts aborting the watching process and restarting it always produces a valid build expected behaviour the build should never fail in the browser only steps to reproduce checkout master run npm i run npm run dev make changes to any file and watch process rebuild reload browser to see resulting changes the browser renders nothing and throws an error uncaught error module build failed from node modules babel loader lib index js error assets src edit story components canvas gridview index js enoent no such file or directory open node modules svgo lib svgo svgo yml the actual file containing the error in the above components canvas gridview index js can vary it can be any file containing a reference to an svg and the file is not related to the change performed triggering the rebuild if no error happened repeat from step until one does occur it takes no more than attempts on my setup screenshots do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria qa instructions
0
109,930
4,415,604,330
IssuesEvent
2016-08-14 06:10:22
DavidLu1997/gogogo
https://api.github.com/repos/DavidLu1997/gogogo
opened
Incorrect GET parameters for game.go
Bug Medium Priority Next Release
Currently, the query string parameter is `user_id` when it should be `game_id`.
1.0
Incorrect GET parameters for game.go - Currently, the query string parameter is `user_id` when it should be `game_id`.
non_process
incorrect get parameters for game go currently the query string parameter is user id when it should be game id
0
2,431
3,845,932,323
IssuesEvent
2016-04-05 00:54:04
brave/browser-laptop
https://api.github.com/repos/brave/browser-laptop
closed
SECURITY: [DID] Installer not timestamped, signed only with SHA1
security upstream
The installer for Windows should probably be dual-signed with SHA256 and SHA1, and should be timestamped to avoid future compatibility problems. http://textslashplain.com/2016/01/10/authenticode-in-2016/
True
SECURITY: [DID] Installer not timestamped, signed only with SHA1 - The installer for Windows should probably be dual-signed with SHA256 and SHA1, and should be timestamped to avoid future compatibility problems. http://textslashplain.com/2016/01/10/authenticode-in-2016/
non_process
security installer not timestamped signed only with the installer for windows should probably be dual signed with and and should be timestamped to avoid future compatibility problems
0
1,064
3,536,062,536
IssuesEvent
2016-01-17 00:14:13
t3kt/vjzual2
https://api.github.com/repos/t3kt/vjzual2
closed
make the displacement slider in the warp module positive-only
video processing
having negative displacement available is probably useful but i almost never use it. so maybe just add "reverse" toggle parameter
1.0
make the displacement slider in the warp module positive-only - having negative displacement available is probably useful but i almost never use it. so maybe just add "reverse" toggle parameter
process
make the displacement slider in the warp module positive only having negative displacement available is probably useful but i almost never use it so maybe just add reverse toggle parameter
1
261,509
19,714,766,384
IssuesEvent
2022-01-13 09:54:45
vanmarcke/ocr_projet6_SnowTricks
https://api.github.com/repos/vanmarcke/ocr_projet6_SnowTricks
opened
README.md update
documentation
**Priority :** 1 **Duration :** 1 **Description:** Update README.md **File added / modified:** **Additional context** Add any other context or screenshots here.
1.0
README.md update - **Priority :** 1 **Duration :** 1 **Description:** Update README.md **File added / modified:** **Additional context** Add any other context or screenshots here.
non_process
readme md update priority duration description update readme md file added modified additional context add any other context or screenshots here
0
8,993
12,103,490,565
IssuesEvent
2020-04-20 18:29:27
GSA/CHRISUpdate
https://api.github.com/repos/GSA/CHRISUpdate
opened
Data Validation: Reject Future Investigation Dates
Bug: Data Bug: Functional Topic: Record Processing Type: Bug
**Discovered By**: CL **Discovery Date**: April 17, 2020 **Discovery Environment**: Production **Severity**: 3 **Description**: The presence of future nac or naci dates on a record produces errors on the manual adjudication page. This issue was discovered because HR Links file process does not check that investigation, nac, and naci dates are <= the current date. This check must be performed in order to avoid this error in the future. **Exception information**: Error Exception: System.Web.HttpUnhandledException (0x80004005): Exception of type 'System.Web.HttpUnhandledException' was thrown. ---> System.FormatException: Input date is invalid at gcims_web.pagebase.GetDbFormatDateString(String dateString) in C:\Users\__________\Source\Repos\GCIMS\gcims_web\pagebase.vb:line 989 at gcims_sa.pmoadjudicationupdate.lstLookupNameList_Changed(Object Sender, EventArgs E) in C:\Users\__________\Source\Repos\GCIMS\gcims_sa\pmoadjudicationupdate.aspx.vb:line 882 at System.Web.UI.WebControls.ListControl.OnSelectedIndexChanged(EventArgs e) at System.Web.UI.Page.RaiseChangedEvents() at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.HandleError(Exception e) at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest() at System.Web.UI.Page.ProcessRequest(HttpContext context) at ASP.sa_pmoadjudicationupdate_aspx.ProcessRequest(HttpContext context) at System.Web.HttpApplication.CallHandlerExecutionStep.System.Web.HttpApplication.IExecutionStep.Execute() at System.Web.HttpApplication.ExecuteStepImpl(IExecutionStep step) at System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously) Error Message: Exception of type 'System.Web.HttpUnhandledException' was thrown. Error Source: System.Web Error Stacktrace: at System.Web.UI.Page.HandleError(Exception e) at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest() at System.Web.UI.Page.ProcessRequest(HttpContext context) at ASP.sa_pmoadjudicationupdate_aspx.ProcessRequest(HttpContext context) at System.Web.HttpApplication.CallHandlerExecutionStep.System.Web.HttpApplication.IExecutionStep.Execute() at System.Web.HttpApplication.ExecuteStepImpl(IExecutionStep step) at System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously) Error TargetSite: Boolean HandleError(System.Exception) Error InnerException: System.FormatException: Input date is invalid at gcims_web.pagebase.GetDbFormatDateString(String dateString) in C:\Users\__________\Source\Repos\GCIMS\gcims_web\pagebase.vb:line 989 at gcims_sa.pmoadjudicationupdate.lstLookupNameList_Changed(Object Sender, EventArgs E) in C:\Users\__________\Source\Repos\GCIMS\gcims_sa\pmoadjudicationupdate.aspx.vb:line 882 at System.Web.UI.WebControls.ListControl.OnSelectedIndexChanged(EventArgs e) at System.Web.UI.Page.RaiseChangedEvents() at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) Error InnerException Message: Input date is invalid Error InnerException Source: System.Web Error InnerException Stacktrace: at System.Web.UI.Page.HandleError(Exception e) at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest() at System.Web.UI.Page.ProcessRequest(HttpContext context) at ASP.sa_pmoadjudicationupdate_aspx.ProcessRequest(HttpContext context) at System.Web.HttpApplication.CallHandlerExecutionStep.System.Web.HttpApplication.IExecutionStep.Execute() at System.Web.HttpApplication.ExecuteStepImpl(IExecutionStep step) at System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously) Error InnerException TargetSite: Boolean HandleError(System.Exception)
1.0
Data Validation: Reject Future Investigation Dates - **Discovered By**: CL **Discovery Date**: April 17, 2020 **Discovery Environment**: Production **Severity**: 3 **Description**: The presence of future nac or naci dates on a record produces errors on the manual adjudication page. This issue was discovered because HR Links file process does not check that investigation, nac, and naci dates are <= the current date. This check must be performed in order to avoid this error in the future. **Exception information**: Error Exception: System.Web.HttpUnhandledException (0x80004005): Exception of type 'System.Web.HttpUnhandledException' was thrown. ---> System.FormatException: Input date is invalid at gcims_web.pagebase.GetDbFormatDateString(String dateString) in C:\Users\__________\Source\Repos\GCIMS\gcims_web\pagebase.vb:line 989 at gcims_sa.pmoadjudicationupdate.lstLookupNameList_Changed(Object Sender, EventArgs E) in C:\Users\__________\Source\Repos\GCIMS\gcims_sa\pmoadjudicationupdate.aspx.vb:line 882 at System.Web.UI.WebControls.ListControl.OnSelectedIndexChanged(EventArgs e) at System.Web.UI.Page.RaiseChangedEvents() at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.HandleError(Exception e) at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest() at System.Web.UI.Page.ProcessRequest(HttpContext context) at ASP.sa_pmoadjudicationupdate_aspx.ProcessRequest(HttpContext context) at System.Web.HttpApplication.CallHandlerExecutionStep.System.Web.HttpApplication.IExecutionStep.Execute() at System.Web.HttpApplication.ExecuteStepImpl(IExecutionStep step) at System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously) Error Message: Exception of type 'System.Web.HttpUnhandledException' was thrown. Error Source: System.Web Error Stacktrace: at System.Web.UI.Page.HandleError(Exception e) at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest() at System.Web.UI.Page.ProcessRequest(HttpContext context) at ASP.sa_pmoadjudicationupdate_aspx.ProcessRequest(HttpContext context) at System.Web.HttpApplication.CallHandlerExecutionStep.System.Web.HttpApplication.IExecutionStep.Execute() at System.Web.HttpApplication.ExecuteStepImpl(IExecutionStep step) at System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously) Error TargetSite: Boolean HandleError(System.Exception) Error InnerException: System.FormatException: Input date is invalid at gcims_web.pagebase.GetDbFormatDateString(String dateString) in C:\Users\__________\Source\Repos\GCIMS\gcims_web\pagebase.vb:line 989 at gcims_sa.pmoadjudicationupdate.lstLookupNameList_Changed(Object Sender, EventArgs E) in C:\Users\__________\Source\Repos\GCIMS\gcims_sa\pmoadjudicationupdate.aspx.vb:line 882 at System.Web.UI.WebControls.ListControl.OnSelectedIndexChanged(EventArgs e) at System.Web.UI.Page.RaiseChangedEvents() at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) Error InnerException Message: Input date is invalid Error InnerException Source: System.Web Error InnerException Stacktrace: at System.Web.UI.Page.HandleError(Exception e) at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint) at System.Web.UI.Page.ProcessRequest() at System.Web.UI.Page.ProcessRequest(HttpContext context) at ASP.sa_pmoadjudicationupdate_aspx.ProcessRequest(HttpContext context) at System.Web.HttpApplication.CallHandlerExecutionStep.System.Web.HttpApplication.IExecutionStep.Execute() at System.Web.HttpApplication.ExecuteStepImpl(IExecutionStep step) at System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously) Error InnerException TargetSite: Boolean HandleError(System.Exception)
process
data validation reject future investigation dates discovered by cl discovery date april discovery environment production severity description the presence of future nac or naci dates on a record produces errors on the manual adjudication page this issue was discovered because hr links file process does not check that investigation nac and naci dates are the current date this check must be performed in order to avoid this error in the future exception information error exception system web httpunhandledexception exception of type system web httpunhandledexception was thrown system formatexception input date is invalid at gcims web pagebase getdbformatdatestring string datestring in c users source repos gcims gcims web pagebase vb line at gcims sa pmoadjudicationupdate lstlookupnamelist changed object sender eventargs e in c users source repos gcims gcims sa pmoadjudicationupdate aspx vb line at system web ui webcontrols listcontrol onselectedindexchanged eventargs e at system web ui page raisechangedevents at system web ui page processrequestmain boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint at system web ui page handleerror exception e at system web ui page processrequestmain boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint at system web ui page processrequest boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint at system web ui page processrequest at system web ui page processrequest httpcontext context at asp sa pmoadjudicationupdate aspx processrequest httpcontext context at system web httpapplication callhandlerexecutionstep system web httpapplication iexecutionstep execute at system web httpapplication executestepimpl iexecutionstep step at system web httpapplication executestep iexecutionstep step boolean completedsynchronously error message exception of type system web httpunhandledexception was thrown error source system web error stacktrace at system web ui page handleerror exception e at system web ui page processrequestmain boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint at system web ui page processrequest boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint at system web ui page processrequest at system web ui page processrequest httpcontext context at asp sa pmoadjudicationupdate aspx processrequest httpcontext context at system web httpapplication callhandlerexecutionstep system web httpapplication iexecutionstep execute at system web httpapplication executestepimpl iexecutionstep step at system web httpapplication executestep iexecutionstep step boolean completedsynchronously error targetsite boolean handleerror system exception error innerexception system formatexception input date is invalid at gcims web pagebase getdbformatdatestring string datestring in c users source repos gcims gcims web pagebase vb line at gcims sa pmoadjudicationupdate lstlookupnamelist changed object sender eventargs e in c users source repos gcims gcims sa pmoadjudicationupdate aspx vb line at system web ui webcontrols listcontrol onselectedindexchanged eventargs e at system web ui page raisechangedevents at system web ui page processrequestmain boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint error innerexception message input date is invalid error innerexception source system web error innerexception stacktrace at system web ui page handleerror exception e at system web ui page processrequestmain boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint at system web ui page processrequest boolean includestagesbeforeasyncpoint boolean includestagesafterasyncpoint at system web ui page processrequest at system web ui page processrequest httpcontext context at asp sa pmoadjudicationupdate aspx processrequest httpcontext context at system web httpapplication callhandlerexecutionstep system web httpapplication iexecutionstep execute at system web httpapplication executestepimpl iexecutionstep step at system web httpapplication executestep iexecutionstep step boolean completedsynchronously error innerexception targetsite boolean handleerror system exception
1
44,763
13,086,695,032
IssuesEvent
2020-08-02 07:45:47
TS-at-WS/cloudify-cli
https://api.github.com/repos/TS-at-WS/cloudify-cli
opened
CVE-2019-16779 (Medium) detected in excon-0.45.4.gem
security vulnerability
## CVE-2019-16779 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>excon-0.45.4.gem</b></p></summary> <p>EXtended http(s) CONnections</p> <p>Library home page: <a href="https://rubygems.org/gems/excon-0.45.4.gem">https://rubygems.org/gems/excon-0.45.4.gem</a></p> <p> Dependency Hierarchy: - kitchen-ec2-0.10.0.gem (Root Library) - :x: **excon-0.45.4.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TS-at-WS/cloudify-cli/commit/598b54ecd67495a76678177f910cdc5eac6128d0">598b54ecd67495a76678177f910cdc5eac6128d0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In RubyGem excon before 0.71.0, there was a race condition around persistent connections, where a connection which is interrupted (such as by a timeout) would leave data on the socket. Subsequent requests would then read this data, returning content from the previous response. The race condition window appears to be short, and it would be difficult to purposefully exploit this. <p>Publish Date: 2019-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16779>CVE-2019-16779</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16779">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16779</a></p> <p>Release Date: 2019-12-16</p> <p>Fix Resolution: excon - 0.71.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Ruby","packageName":"excon","packageVersion":"0.45.4","isTransitiveDependency":true,"dependencyTree":"kitchen-ec2:0.10.0;excon:0.45.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"excon - 0.71.0"}],"vulnerabilityIdentifier":"CVE-2019-16779","vulnerabilityDetails":"In RubyGem excon before 0.71.0, there was a race condition around persistent connections, where a connection which is interrupted (such as by a timeout) would leave data on the socket. Subsequent requests would then read this data, returning content from the previous response. The race condition window appears to be short, and it would be difficult to purposefully exploit this.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16779","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-16779 (Medium) detected in excon-0.45.4.gem - ## CVE-2019-16779 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>excon-0.45.4.gem</b></p></summary> <p>EXtended http(s) CONnections</p> <p>Library home page: <a href="https://rubygems.org/gems/excon-0.45.4.gem">https://rubygems.org/gems/excon-0.45.4.gem</a></p> <p> Dependency Hierarchy: - kitchen-ec2-0.10.0.gem (Root Library) - :x: **excon-0.45.4.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TS-at-WS/cloudify-cli/commit/598b54ecd67495a76678177f910cdc5eac6128d0">598b54ecd67495a76678177f910cdc5eac6128d0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In RubyGem excon before 0.71.0, there was a race condition around persistent connections, where a connection which is interrupted (such as by a timeout) would leave data on the socket. Subsequent requests would then read this data, returning content from the previous response. The race condition window appears to be short, and it would be difficult to purposefully exploit this. <p>Publish Date: 2019-12-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16779>CVE-2019-16779</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16779">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16779</a></p> <p>Release Date: 2019-12-16</p> <p>Fix Resolution: excon - 0.71.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Ruby","packageName":"excon","packageVersion":"0.45.4","isTransitiveDependency":true,"dependencyTree":"kitchen-ec2:0.10.0;excon:0.45.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"excon - 0.71.0"}],"vulnerabilityIdentifier":"CVE-2019-16779","vulnerabilityDetails":"In RubyGem excon before 0.71.0, there was a race condition around persistent connections, where a connection which is interrupted (such as by a timeout) would leave data on the socket. Subsequent requests would then read this data, returning content from the previous response. The race condition window appears to be short, and it would be difficult to purposefully exploit this.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16779","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in excon gem cve medium severity vulnerability vulnerable library excon gem extended http s connections library home page a href dependency hierarchy kitchen gem root library x excon gem vulnerable library found in head commit a href vulnerability details in rubygem excon before there was a race condition around persistent connections where a connection which is interrupted such as by a timeout would leave data on the socket subsequent requests would then read this data returning content from the previous response the race condition window appears to be short and it would be difficult to purposefully exploit this publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution excon isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in rubygem excon before there was a race condition around persistent connections where a connection which is interrupted such as by a timeout would leave data on the socket subsequent requests would then read this data returning content from the previous response the race condition window appears to be short and it would be difficult to purposefully exploit this vulnerabilityurl
0
18,292
24,399,439,275
IssuesEvent
2022-10-04 23:01:37
python/cpython
https://api.github.com/repos/python/cpython
closed
multiprocessing AF_PIPE name format is slightly confusing in the docs
type-feature docs 3.11 3.10 3.9 3.8 expert-multiprocessing
BPO | [44189](https://bugs.python.org/issue44189) --- | :--- Nosy | @anntzer, @eryksun <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2021-05-20.12:11:12.410> labels = ['3.8', '3.9', '3.10', '3.11', 'type-feature', 'docs'] title = 'multiprocessing AF_PIPE name format is slightly confusing in the docs' updated_at = <Date 2021-05-20.23:19:26.845> user = 'https://github.com/anntzer' ``` bugs.python.org fields: ```python activity = <Date 2021-05-20.23:19:26.845> actor = 'eryksun' assignee = 'docs@python' closed = False closed_date = None closer = None components = ['Documentation'] creation = <Date 2021-05-20.12:11:12.410> creator = 'Antony.Lee' dependencies = [] files = [] hgrepos = [] issue_num = 44189 keywords = [] message_count = 2.0 messages = ['394020', '394078'] nosy_count = 3.0 nosy_names = ['docs@python', 'Antony.Lee', 'eryksun'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'enhancement' url = 'https://bugs.python.org/issue44189' versions = ['Python 3.8', 'Python 3.9', 'Python 3.10', 'Python 3.11'] ``` </p></details>
1.0
multiprocessing AF_PIPE name format is slightly confusing in the docs - BPO | [44189](https://bugs.python.org/issue44189) --- | :--- Nosy | @anntzer, @eryksun <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2021-05-20.12:11:12.410> labels = ['3.8', '3.9', '3.10', '3.11', 'type-feature', 'docs'] title = 'multiprocessing AF_PIPE name format is slightly confusing in the docs' updated_at = <Date 2021-05-20.23:19:26.845> user = 'https://github.com/anntzer' ``` bugs.python.org fields: ```python activity = <Date 2021-05-20.23:19:26.845> actor = 'eryksun' assignee = 'docs@python' closed = False closed_date = None closer = None components = ['Documentation'] creation = <Date 2021-05-20.12:11:12.410> creator = 'Antony.Lee' dependencies = [] files = [] hgrepos = [] issue_num = 44189 keywords = [] message_count = 2.0 messages = ['394020', '394078'] nosy_count = 3.0 nosy_names = ['docs@python', 'Antony.Lee', 'eryksun'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = 'enhancement' url = 'https://bugs.python.org/issue44189' versions = ['Python 3.8', 'Python 3.9', 'Python 3.10', 'Python 3.11'] ``` </p></details>
process
multiprocessing af pipe name format is slightly confusing in the docs bpo nosy anntzer eryksun note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title multiprocessing af pipe name format is slightly confusing in the docs updated at user bugs python org fields python activity actor eryksun assignee docs python closed false closed date none closer none components creation creator antony lee dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type enhancement url versions
1
336,081
10,171,167,004
IssuesEvent
2019-08-08 07:43:22
laterpay/laterpay-wordpress-plugin
https://api.github.com/repos/laterpay/laterpay-wordpress-plugin
closed
Allow merchants to Opt in or Out of Wisdom data capture on Settings Page
Low Priority
# User Story As a merchant, I would like to be able to change my mind later in case that I develop security concerns or learn to trust LaterPay after more experience with their product. As a LaterPay support person, I would like to be able to request that a merchant turn on tracking so that I can provide better support. # Acceptance Criteria - Allow merchants to Opt in or Out after Activation on the LaterPay Settings Page ([here](https://wisdomplugin.com/support/#advanced)) - Update the existing Analytics section, as follows ### Intro Paragraph **English** > In order to make informed decisions about improvements to our plugin, we would like to track its usage on your site. We do not record any sensitive information and are not in the business of selling data. The only metrics that we track are in regards to your WordPress environment and plugin settings. If you would like to add LaterPay events to your own Google Analytics environment, simply link to your Google Analytics account in the box below. Additional analytics on your site's purchase history are available in your LaterPay Merchant Analytics Dashboard. To view this information, simply log in to your LaterPay account on [laterpay.net](https://www.laterpay.net/). **German** > Um fundierte Entscheidungen über Verbesserungen an unserem Plugin zu treffen, möchten wir dessen Verwendung auf Ihrer Website verfolgen. Wir erfassen keine sensiblen Informationen und sind nicht im Geschäft mit dem Verkauf von Daten. Die einzigen Kennzahlen, die wir verfolgen, beziehen sich auf Ihre WordPress-Umgebung und die Plugin-Einstellungen. Wenn Sie LaterPay-Ereignisse zu Ihrer eigenen Google Analytics-Umgebung hinzufügen möchten, verlinken Sie einfach auf Ihr Google Analytics-Konto in der untenstehenden Box. Zusätzliche Analysen zur Kaufhistorie Ihrer Website sind in Ihrem LaterPay Merchant Analytics Dashboard verfügbar. Um diese Informationen einzusehen, loggen Sie sich einfach in Ihr LaterPay-Konto auf [laterpay.net](https://www.laterpay.net/) ein. ### New Checkbox **English** > Track Plugin Usage & WordPress Settings: **German** > Verfolgen Sie die Verwendung von Plugins und WordPress-Einstellungen: # Mockup ![image](https://user-images.githubusercontent.com/41650164/60190166-ccc9bc00-97f7-11e9-8312-af43d4859509.png)
1.0
Allow merchants to Opt in or Out of Wisdom data capture on Settings Page - # User Story As a merchant, I would like to be able to change my mind later in case that I develop security concerns or learn to trust LaterPay after more experience with their product. As a LaterPay support person, I would like to be able to request that a merchant turn on tracking so that I can provide better support. # Acceptance Criteria - Allow merchants to Opt in or Out after Activation on the LaterPay Settings Page ([here](https://wisdomplugin.com/support/#advanced)) - Update the existing Analytics section, as follows ### Intro Paragraph **English** > In order to make informed decisions about improvements to our plugin, we would like to track its usage on your site. We do not record any sensitive information and are not in the business of selling data. The only metrics that we track are in regards to your WordPress environment and plugin settings. If you would like to add LaterPay events to your own Google Analytics environment, simply link to your Google Analytics account in the box below. Additional analytics on your site's purchase history are available in your LaterPay Merchant Analytics Dashboard. To view this information, simply log in to your LaterPay account on [laterpay.net](https://www.laterpay.net/). **German** > Um fundierte Entscheidungen über Verbesserungen an unserem Plugin zu treffen, möchten wir dessen Verwendung auf Ihrer Website verfolgen. Wir erfassen keine sensiblen Informationen und sind nicht im Geschäft mit dem Verkauf von Daten. Die einzigen Kennzahlen, die wir verfolgen, beziehen sich auf Ihre WordPress-Umgebung und die Plugin-Einstellungen. Wenn Sie LaterPay-Ereignisse zu Ihrer eigenen Google Analytics-Umgebung hinzufügen möchten, verlinken Sie einfach auf Ihr Google Analytics-Konto in der untenstehenden Box. Zusätzliche Analysen zur Kaufhistorie Ihrer Website sind in Ihrem LaterPay Merchant Analytics Dashboard verfügbar. Um diese Informationen einzusehen, loggen Sie sich einfach in Ihr LaterPay-Konto auf [laterpay.net](https://www.laterpay.net/) ein. ### New Checkbox **English** > Track Plugin Usage & WordPress Settings: **German** > Verfolgen Sie die Verwendung von Plugins und WordPress-Einstellungen: # Mockup ![image](https://user-images.githubusercontent.com/41650164/60190166-ccc9bc00-97f7-11e9-8312-af43d4859509.png)
non_process
allow merchants to opt in or out of wisdom data capture on settings page user story as a merchant i would like to be able to change my mind later in case that i develop security concerns or learn to trust laterpay after more experience with their product as a laterpay support person i would like to be able to request that a merchant turn on tracking so that i can provide better support acceptance criteria allow merchants to opt in or out after activation on the laterpay settings page update the existing analytics section as follows intro paragraph english in order to make informed decisions about improvements to our plugin we would like to track its usage on your site we do not record any sensitive information and are not in the business of selling data the only metrics that we track are in regards to your wordpress environment and plugin settings if you would like to add laterpay events to your own google analytics environment simply link to your google analytics account in the box below additional analytics on your site s purchase history are available in your laterpay merchant analytics dashboard to view this information simply log in to your laterpay account on german um fundierte entscheidungen über verbesserungen an unserem plugin zu treffen möchten wir dessen verwendung auf ihrer website verfolgen wir erfassen keine sensiblen informationen und sind nicht im geschäft mit dem verkauf von daten die einzigen kennzahlen die wir verfolgen beziehen sich auf ihre wordpress umgebung und die plugin einstellungen wenn sie laterpay ereignisse zu ihrer eigenen google analytics umgebung hinzufügen möchten verlinken sie einfach auf ihr google analytics konto in der untenstehenden box zusätzliche analysen zur kaufhistorie ihrer website sind in ihrem laterpay merchant analytics dashboard verfügbar um diese informationen einzusehen loggen sie sich einfach in ihr laterpay konto auf ein new checkbox english track plugin usage wordpress settings german verfolgen sie die verwendung von plugins und wordpress einstellungen mockup
0
4,752
7,612,069,255
IssuesEvent
2018-05-01 16:11:11
googlegenomics/gcp-variant-transforms
https://api.github.com/repos/googlegenomics/gcp-variant-transforms
opened
Change docker run in README to use pipelines v2
P2 process
The existing docker run command uses pipelines v1; we should update it to use v2 since gcloud support is now available.
1.0
Change docker run in README to use pipelines v2 - The existing docker run command uses pipelines v1; we should update it to use v2 since gcloud support is now available.
process
change docker run in readme to use pipelines the existing docker run command uses pipelines we should update it to use since gcloud support is now available
1