Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
10,232
2,935,661,500
IssuesEvent
2015-06-30 15:26:27
kleingeist/muster
https://api.github.com/repos/kleingeist/muster
closed
Layout: Favoritenseite
design
Idee: angelehnt oder kombiniert mit Suchseite (Favorit ist ein Suchmerkmal)
1.0
Layout: Favoritenseite - Idee: angelehnt oder kombiniert mit Suchseite (Favorit ist ein Suchmerkmal)
non_process
layout favoritenseite idee angelehnt oder kombiniert mit suchseite favorit ist ein suchmerkmal
0
523,367
15,179,291,640
IssuesEvent
2021-02-14 18:56:30
apcountryman/picolibrary
https://api.github.com/repos/apcountryman/picolibrary
closed
Add I2C interactive testing facilities namespace
priority-normal status-awaiting_approval type-feature
Add I2C interactive testing facilities namespace (`::picolibrary::Testing::Interactive::I2C`) and the associated header/source file pair (`include/picolibrary/testing/interactive/i2.h` and `source/picolibrary/testing/interactive/i2c.cc`)
1.0
Add I2C interactive testing facilities namespace - Add I2C interactive testing facilities namespace (`::picolibrary::Testing::Interactive::I2C`) and the associated header/source file pair (`include/picolibrary/testing/interactive/i2.h` and `source/picolibrary/testing/interactive/i2c.cc`)
non_process
add interactive testing facilities namespace add interactive testing facilities namespace picolibrary testing interactive and the associated header source file pair include picolibrary testing interactive h and source picolibrary testing interactive cc
0
3,448
6,541,387,467
IssuesEvent
2017-09-01 19:42:24
amaster507/ifbmt
https://api.github.com/repos/amaster507/ifbmt
opened
Schedule Booking Confirmation
email idea process
When booking an appointment, there needs to be a process to send confirmation to the email address on file or an updated email if provided. This confirmation should send to the contact at the church and Carbon Copy the missionary for record keeping and confirmation.
1.0
Schedule Booking Confirmation - When booking an appointment, there needs to be a process to send confirmation to the email address on file or an updated email if provided. This confirmation should send to the contact at the church and Carbon Copy the missionary for record keeping and confirmation.
process
schedule booking confirmation when booking an appointment there needs to be a process to send confirmation to the email address on file or an updated email if provided this confirmation should send to the contact at the church and carbon copy the missionary for record keeping and confirmation
1
303,174
22,957,737,307
IssuesEvent
2022-07-19 13:04:07
cloudflare/cloudflare-docs
https://api.github.com/repos/cloudflare/cloudflare-docs
closed
[Pages] SvelteKit Adapter for Pages
documentation content:edit
### Which Cloudflare product does this pertain to? Pages ### Existing documentation URL(s) https://developers.cloudflare.com/pages/framework-guides/deploy-a-svelte-site/#deploying-with-cloudflare-pages ### Section that requires update "Deploying with Cloudflare Pages" and its subsection "SvelteKit Configuration" ### What needs to change? The documentation suggests us to use `adapter-static` SvelteKit adapter, but I suppose this should be updated since [`adapter-cloudflare`](https://www.npmjs.com/package/@sveltejs/adapter-cloudflare) has been introduced. Quote from `adapter-cloudflare`'s README: > - adapter-cloudflare – supports all SvelteKit features; builds for Cloudflare Pages > - adapter-cloudflare-workers – supports all SvelteKit features; builds for Cloudflare Workers > - adapter-static – only produces client-side static assets; compatible with Cloudflare Pages > > Note: Cloudflare Pages' new Workers integration is currently in beta. > Compared to adapter-cloudflare-workers, this adapter will be the preferred approach for most users since building on top of Pages unlocks automatic builds and deploys, preview deployments, instant rollbacks, etc. > From SvelteKit's perspective, there is no difference and no functionality loss when migrating to/from the Workers and the Pages adapters. ### How should it change? Since `adapter-cloudflare` is a built-in feature for the default adapter, `adapter-auto`, this subsection "SvelteKit Configuration" could be deleted entirely. As for the build parameters mentioned in the section "Deploying with Cloudflare Pages", the build (output) directory should be `.svelte-kit/cloudflare/` and `NODE_VERSION` should be set to `16` if you are going to use later SvelteKit (https://github.com/sveltejs/kit/issues/3003#issuecomment-1059609542). ### Additional information _No response_
1.0
[Pages] SvelteKit Adapter for Pages - ### Which Cloudflare product does this pertain to? Pages ### Existing documentation URL(s) https://developers.cloudflare.com/pages/framework-guides/deploy-a-svelte-site/#deploying-with-cloudflare-pages ### Section that requires update "Deploying with Cloudflare Pages" and its subsection "SvelteKit Configuration" ### What needs to change? The documentation suggests us to use `adapter-static` SvelteKit adapter, but I suppose this should be updated since [`adapter-cloudflare`](https://www.npmjs.com/package/@sveltejs/adapter-cloudflare) has been introduced. Quote from `adapter-cloudflare`'s README: > - adapter-cloudflare – supports all SvelteKit features; builds for Cloudflare Pages > - adapter-cloudflare-workers – supports all SvelteKit features; builds for Cloudflare Workers > - adapter-static – only produces client-side static assets; compatible with Cloudflare Pages > > Note: Cloudflare Pages' new Workers integration is currently in beta. > Compared to adapter-cloudflare-workers, this adapter will be the preferred approach for most users since building on top of Pages unlocks automatic builds and deploys, preview deployments, instant rollbacks, etc. > From SvelteKit's perspective, there is no difference and no functionality loss when migrating to/from the Workers and the Pages adapters. ### How should it change? Since `adapter-cloudflare` is a built-in feature for the default adapter, `adapter-auto`, this subsection "SvelteKit Configuration" could be deleted entirely. As for the build parameters mentioned in the section "Deploying with Cloudflare Pages", the build (output) directory should be `.svelte-kit/cloudflare/` and `NODE_VERSION` should be set to `16` if you are going to use later SvelteKit (https://github.com/sveltejs/kit/issues/3003#issuecomment-1059609542). ### Additional information _No response_
non_process
sveltekit adapter for pages which cloudflare product does this pertain to pages existing documentation url s section that requires update deploying with cloudflare pages and its subsection sveltekit configuration what needs to change the documentation suggests us to use adapter static sveltekit adapter but i suppose this should be updated since has been introduced quote from adapter cloudflare s readme adapter cloudflare – supports all sveltekit features builds for cloudflare pages adapter cloudflare workers – supports all sveltekit features builds for cloudflare workers adapter static – only produces client side static assets compatible with cloudflare pages note cloudflare pages new workers integration is currently in beta compared to adapter cloudflare workers this adapter will be the preferred approach for most users since building on top of pages unlocks automatic builds and deploys preview deployments instant rollbacks etc from sveltekit s perspective there is no difference and no functionality loss when migrating to from the workers and the pages adapters how should it change since adapter cloudflare is a built in feature for the default adapter adapter auto this subsection sveltekit configuration could be deleted entirely as for the build parameters mentioned in the section deploying with cloudflare pages the build output directory should be svelte kit cloudflare and node version should be set to if you are going to use later sveltekit additional information no response
0
14,642
17,772,751,883
IssuesEvent
2021-08-30 15:23:28
trixi-framework/Trixi.jl
https://api.github.com/repos/trixi-framework/Trixi.jl
closed
Where to put postprocessing/visualization routines?
question postprocessing visualization
@sloede in https://github.com/trixi-framework/Trixi.jl/pull/403#issue-543686261: > In general, I am currently reconsidering whether the decision to pull the plotting functionality into the Trixi package was the right decision. Maybe instead of consolidating these auxiliary methods in the main package, we should consider creating a "postprocessing package" (TrixiPostprocessing?) that consolidates things such as the plot recipes, interpolation methods, plot methods, VTK writing etc.? But maybe this is something that does not to be (or even can be) answered right now but that we should get the functionality first and consider these structural decisions separately in a refactoring step. @ranocha in https://github.com/trixi-framework/Trixi.jl/pull/403#issuecomment-749359325: > The advantage of having the plot recipes in Trixi.jl is that users don't need to load another package just to be able to plot(sol) - implementing the plot recipes in another package would also be type piracy. Of course, the situation is different for things like additional callbacks, VTK stuff etc. Your suggestion to get the features first, mark them as experimental, and see where they should live after working with them sounds good to me.
1.0
Where to put postprocessing/visualization routines? - @sloede in https://github.com/trixi-framework/Trixi.jl/pull/403#issue-543686261: > In general, I am currently reconsidering whether the decision to pull the plotting functionality into the Trixi package was the right decision. Maybe instead of consolidating these auxiliary methods in the main package, we should consider creating a "postprocessing package" (TrixiPostprocessing?) that consolidates things such as the plot recipes, interpolation methods, plot methods, VTK writing etc.? But maybe this is something that does not to be (or even can be) answered right now but that we should get the functionality first and consider these structural decisions separately in a refactoring step. @ranocha in https://github.com/trixi-framework/Trixi.jl/pull/403#issuecomment-749359325: > The advantage of having the plot recipes in Trixi.jl is that users don't need to load another package just to be able to plot(sol) - implementing the plot recipes in another package would also be type piracy. Of course, the situation is different for things like additional callbacks, VTK stuff etc. Your suggestion to get the features first, mark them as experimental, and see where they should live after working with them sounds good to me.
process
where to put postprocessing visualization routines sloede in in general i am currently reconsidering whether the decision to pull the plotting functionality into the trixi package was the right decision maybe instead of consolidating these auxiliary methods in the main package we should consider creating a postprocessing package trixipostprocessing that consolidates things such as the plot recipes interpolation methods plot methods vtk writing etc but maybe this is something that does not to be or even can be answered right now but that we should get the functionality first and consider these structural decisions separately in a refactoring step ranocha in the advantage of having the plot recipes in trixi jl is that users don t need to load another package just to be able to plot sol implementing the plot recipes in another package would also be type piracy of course the situation is different for things like additional callbacks vtk stuff etc your suggestion to get the features first mark them as experimental and see where they should live after working with them sounds good to me
1
640,648
20,795,300,756
IssuesEvent
2022-03-17 08:41:56
COS-301/graduates
https://api.github.com/repos/COS-301/graduates
opened
API engineer
priority:high role:api-engineer
- [ ] Design how the api layer will communicate with the service layer, how data will be accessed - [ ] Mock code for requesting data from the database (receiving requests from service and requesting that data from the database)
1.0
API engineer - - [ ] Design how the api layer will communicate with the service layer, how data will be accessed - [ ] Mock code for requesting data from the database (receiving requests from service and requesting that data from the database)
non_process
api engineer design how the api layer will communicate with the service layer how data will be accessed mock code for requesting data from the database receiving requests from service and requesting that data from the database
0
5,182
4,819,895,003
IssuesEvent
2016-11-04 20:47:15
rweda/gaddomnit
https://api.github.com/repos/rweda/gaddomnit
closed
Skip evaluating styles for known invisible tags
enhancement output-size performance
`head, title, link, meta, style, script` do not appear (in most documents), so include an option to skip evaluating their computed style all-together. ~~Investigate if SVG `defs` and it's children can also be unstyled, and have styles applied to their equivalent `use` statements.~~
True
Skip evaluating styles for known invisible tags - `head, title, link, meta, style, script` do not appear (in most documents), so include an option to skip evaluating their computed style all-together. ~~Investigate if SVG `defs` and it's children can also be unstyled, and have styles applied to their equivalent `use` statements.~~
non_process
skip evaluating styles for known invisible tags head title link meta style script do not appear in most documents so include an option to skip evaluating their computed style all together investigate if svg defs and it s children can also be unstyled and have styles applied to their equivalent use statements
0
0
2,490,131,154
IssuesEvent
2015-01-02 08:34:53
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
InetAddress blacklist filter broken
bug processing
Since removing the remote ip lookup blunder (where each message triggered a DNS lookup) we've removed the inetaddress field from the message object. Since the transport/codec refactoring we are now saving the remote ip for socket based transports by default, being careful not to trigger dns lookups at all. This means we can enable the blacklist filter again, by converting the remote ip string back to a InetAddress for blacklisting.
1.0
InetAddress blacklist filter broken - Since removing the remote ip lookup blunder (where each message triggered a DNS lookup) we've removed the inetaddress field from the message object. Since the transport/codec refactoring we are now saving the remote ip for socket based transports by default, being careful not to trigger dns lookups at all. This means we can enable the blacklist filter again, by converting the remote ip string back to a InetAddress for blacklisting.
process
inetaddress blacklist filter broken since removing the remote ip lookup blunder where each message triggered a dns lookup we ve removed the inetaddress field from the message object since the transport codec refactoring we are now saving the remote ip for socket based transports by default being careful not to trigger dns lookups at all this means we can enable the blacklist filter again by converting the remote ip string back to a inetaddress for blacklisting
1
16,579
21,610,192,189
IssuesEvent
2022-05-04 09:18:13
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Extend the `.GetCorrelationInfo` Service Bus message extension to allow operation parent ID extraction
area:correlation area:message-processing
**Is your feature request related to a problem? Please describe.** Currently, we are able to generate a `MessageCorelationInfo` model from a received Azure Service Bus message with an operation ID and transaction ID. But we aren't able to extract the operation parent ID from the `UserProperties` like the others. **Describe the solution you'd like** Provide a way to extract the operation parent ID from the received Azure Service Bus message with as default user property name: `Operation-Parent-Id`. **Additional context** Related to PoC: https://github.com/arcus-azure/arcus-service-to-service-correlation-poc
1.0
Extend the `.GetCorrelationInfo` Service Bus message extension to allow operation parent ID extraction - **Is your feature request related to a problem? Please describe.** Currently, we are able to generate a `MessageCorelationInfo` model from a received Azure Service Bus message with an operation ID and transaction ID. But we aren't able to extract the operation parent ID from the `UserProperties` like the others. **Describe the solution you'd like** Provide a way to extract the operation parent ID from the received Azure Service Bus message with as default user property name: `Operation-Parent-Id`. **Additional context** Related to PoC: https://github.com/arcus-azure/arcus-service-to-service-correlation-poc
process
extend the getcorrelationinfo service bus message extension to allow operation parent id extraction is your feature request related to a problem please describe currently we are able to generate a messagecorelationinfo model from a received azure service bus message with an operation id and transaction id but we aren t able to extract the operation parent id from the userproperties like the others describe the solution you d like provide a way to extract the operation parent id from the received azure service bus message with as default user property name operation parent id additional context related to poc
1
11,264
14,050,208,178
IssuesEvent
2020-11-02 11:23:49
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Add DXF export functionality to processing
Feature Request Processing
Author Name: **Nyall Dawson** (@nyalldawson) Original Redmine Issue: [17495](https://issues.qgis.org/issues/17495) Redmine category:processing/qgis Assignee: Victor Olaya --- Currently DXF export is only available from the project menu -> "DXF export" action. We should expose this as a processing algorithm so that it can be utilised by models, and then just call the processing algorithm from the Project Menu action.
1.0
Add DXF export functionality to processing - Author Name: **Nyall Dawson** (@nyalldawson) Original Redmine Issue: [17495](https://issues.qgis.org/issues/17495) Redmine category:processing/qgis Assignee: Victor Olaya --- Currently DXF export is only available from the project menu -> "DXF export" action. We should expose this as a processing algorithm so that it can be utilised by models, and then just call the processing algorithm from the Project Menu action.
process
add dxf export functionality to processing author name nyall dawson nyalldawson original redmine issue redmine category processing qgis assignee victor olaya currently dxf export is only available from the project menu dxf export action we should expose this as a processing algorithm so that it can be utilised by models and then just call the processing algorithm from the project menu action
1
32,557
4,776,226,264
IssuesEvent
2016-10-27 13:10:23
AlexZH1979/tutorials-3
https://api.github.com/repos/AlexZH1979/tutorials-3
opened
Tutorial Page note.md Issue
Meta tutorials/test/note.md
Tutorial issue found: [https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/note.md](https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/note.md) is missing meta or contains invalid front matter. Your tutorial in sapcom was not updated.
1.0
Tutorial Page note.md Issue - Tutorial issue found: [https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/note.md](https://github.com/AlexZH1979/tutorials-3/blob/master/tutorials/test/note.md) is missing meta or contains invalid front matter. Your tutorial in sapcom was not updated.
non_process
tutorial page note md issue tutorial issue found is missing meta or contains invalid front matter your tutorial in sapcom was not updated
0
12,804
15,181,562,602
IssuesEvent
2021-02-15 03:59:12
Geonovum/disgeo-arch
https://api.github.com/repos/Geonovum/disgeo-arch
closed
5.2.1. Registratie component(en) - a.u.b. niet monoliet, maar klein en flexibel
Flexibiliteit In Behandeling In behandeling - voorstel processen e.d. Plaatjes lagen Processen Functies Componenten
Ik zou het verschrikkelijk vinden, om allerlei redenen, als de SOR fysiek in 1 nieuwe registratie geïmplementeerd worden (kapitaalvernietiging, niet te beheren). Er is 1 ding waar ik bang voor ben, en dat is een monolithische uitwerking van de SOR. Niet dat dit zo wordt bedacht, maar ik lees ook nergens dat een monoliet expliciet buiten scope is als alternatief. De redenen hiervoor: * Enorme kapitaalvernietiging van wat er door de jaren heen gemaakt is. * De trend tegenwoordig en een oplossing die wendbaarheid biedt, is toch echt: maak kleine losely coupled componenten die apart van elkaar aangepast kunnen worden. Dus eerder meer kleinere basisregistraties dan minder, eerder meer informatiemodellen (of losjes gekoppelde onderdelen in een groter ding die de samenhang bewaakt) dan minder enz. - dat gegevensmodellen makkelijk aan te passen zijn en dat informatie meegroeit met de behoefte en in samenhang bruikbaar is, dan juist is het nodig om met dynamiek om te kunnen gaan. Een monoliet past daar uitermate slecht bij. * Een monoliet - 1 registratie component - lost de problemen en uitdagingen niet op Samenhang is een concept die ik meer zie als: in samenhang inwinnen, in samenhang gebruiken, maar niet als: samenbrengen in 1 grote IT component. Dat zou voor mij de bietenbrug zijn… goed, ik weet niet waar het heen beweegt, maar ik krijg wel soms wat monolithische zinnen voor ogen. Wellicht door de ‘bang voor’ bril ingegeven. Als ik dan kijk naar de volgende paragraaf, dan lees ik nergens dat een monoliet uitgesloten wordt. De crux zit 'm in: hoe laat je de registatie componenten zodanig acteren dat deze fijn data ontsluiten voor gebruik. Dat is echt DE design uitdaging voor de SOR. > 5.2.1 Registratie > De component Registratie heeft als doel om bronhouderorganisaties en gemachtigde organisaties in staat te stellen objectgegevens en bijbehorende meta-gegevens te beheren (toevoegen en wijzigen). Deze component biedt de services die bronhouders en gemachtigden daarvoor nodig hebben. >
2.0
5.2.1. Registratie component(en) - a.u.b. niet monoliet, maar klein en flexibel - Ik zou het verschrikkelijk vinden, om allerlei redenen, als de SOR fysiek in 1 nieuwe registratie geïmplementeerd worden (kapitaalvernietiging, niet te beheren). Er is 1 ding waar ik bang voor ben, en dat is een monolithische uitwerking van de SOR. Niet dat dit zo wordt bedacht, maar ik lees ook nergens dat een monoliet expliciet buiten scope is als alternatief. De redenen hiervoor: * Enorme kapitaalvernietiging van wat er door de jaren heen gemaakt is. * De trend tegenwoordig en een oplossing die wendbaarheid biedt, is toch echt: maak kleine losely coupled componenten die apart van elkaar aangepast kunnen worden. Dus eerder meer kleinere basisregistraties dan minder, eerder meer informatiemodellen (of losjes gekoppelde onderdelen in een groter ding die de samenhang bewaakt) dan minder enz. - dat gegevensmodellen makkelijk aan te passen zijn en dat informatie meegroeit met de behoefte en in samenhang bruikbaar is, dan juist is het nodig om met dynamiek om te kunnen gaan. Een monoliet past daar uitermate slecht bij. * Een monoliet - 1 registratie component - lost de problemen en uitdagingen niet op Samenhang is een concept die ik meer zie als: in samenhang inwinnen, in samenhang gebruiken, maar niet als: samenbrengen in 1 grote IT component. Dat zou voor mij de bietenbrug zijn… goed, ik weet niet waar het heen beweegt, maar ik krijg wel soms wat monolithische zinnen voor ogen. Wellicht door de ‘bang voor’ bril ingegeven. Als ik dan kijk naar de volgende paragraaf, dan lees ik nergens dat een monoliet uitgesloten wordt. De crux zit 'm in: hoe laat je de registatie componenten zodanig acteren dat deze fijn data ontsluiten voor gebruik. Dat is echt DE design uitdaging voor de SOR. > 5.2.1 Registratie > De component Registratie heeft als doel om bronhouderorganisaties en gemachtigde organisaties in staat te stellen objectgegevens en bijbehorende meta-gegevens te beheren (toevoegen en wijzigen). Deze component biedt de services die bronhouders en gemachtigden daarvoor nodig hebben. >
process
registratie component en a u b niet monoliet maar klein en flexibel ik zou het verschrikkelijk vinden om allerlei redenen als de sor fysiek in nieuwe registratie geïmplementeerd worden kapitaalvernietiging niet te beheren er is ding waar ik bang voor ben en dat is een monolithische uitwerking van de sor niet dat dit zo wordt bedacht maar ik lees ook nergens dat een monoliet expliciet buiten scope is als alternatief de redenen hiervoor enorme kapitaalvernietiging van wat er door de jaren heen gemaakt is de trend tegenwoordig en een oplossing die wendbaarheid biedt is toch echt maak kleine losely coupled componenten die apart van elkaar aangepast kunnen worden dus eerder meer kleinere basisregistraties dan minder eerder meer informatiemodellen of losjes gekoppelde onderdelen in een groter ding die de samenhang bewaakt dan minder enz dat gegevensmodellen makkelijk aan te passen zijn en dat informatie meegroeit met de behoefte en in samenhang bruikbaar is dan juist is het nodig om met dynamiek om te kunnen gaan een monoliet past daar uitermate slecht bij een monoliet registratie component lost de problemen en uitdagingen niet op samenhang is een concept die ik meer zie als in samenhang inwinnen in samenhang gebruiken maar niet als samenbrengen in grote it component dat zou voor mij de bietenbrug zijn… goed ik weet niet waar het heen beweegt maar ik krijg wel soms wat monolithische zinnen voor ogen wellicht door de ‘bang voor’ bril ingegeven als ik dan kijk naar de volgende paragraaf dan lees ik nergens dat een monoliet uitgesloten wordt de crux zit m in hoe laat je de registatie componenten zodanig acteren dat deze fijn data ontsluiten voor gebruik dat is echt de design uitdaging voor de sor registratie de component registratie heeft als doel om bronhouderorganisaties en gemachtigde organisaties in staat te stellen objectgegevens en bijbehorende meta gegevens te beheren toevoegen en wijzigen deze component biedt de services die bronhouders en gemachtigden daarvoor nodig hebben
1
4,230
7,182,674,894
IssuesEvent
2018-02-01 10:35:51
GeographicaGS/AquaGIS
https://api.github.com/repos/GeographicaGS/AquaGIS
closed
Motor detección de fugas
processing
Lanzamos una tarea en el processing que llame a una función plpgsql, aplique las reglas de fuga para cada sector y actualice la tabla de resultados de fugas. ### Reglas de fuga En una tabla tenemos la configuración de reglas de las fugas "leak_rules" . Algo como: q | p |  time | status 20 | null | 3600 |  1 40 |  -40 | 7200 | 2 status=1: Anomalía de consumo status=2: Peligro de fuga La tabla de last_data de los sectores debe tener una columna que sea "leak_status" y que tenga actualizado el último estado. Si es ok el valor debe ser 0. ### Tabla de resultados Tenemos una tabla con el histórico de leaks por sectores. "leak_sector_historic". Esta tabla se actualiza cuando una fuga cambia de estado. sector_id  | TimeInstant | status | rule Esta tabla guarda los cambios de estado de los sectores. Cuando hay un cambio de estado se crea un nuevo registro para el nuevo estado. Ejemplo: 1. Se detecta una anomalía de consumo. Se añade este registro: "1;2017-01-01T00:00Z;1;'Se detecta un consumo superior al 20%'" 2. Se detecta a la hora siguiente que la anomalía se ha corregido. Se actualiza el registro anterior "1;2017-01-01T00:00Z;1;'Se detecta un consumo superior al 20%'" Se añade un nuevo registro "1;2017-01-01T01:00Z;0;null" Hay que generar una variable Urbo para el campo status de esta tabla. La generación de tablas hay que meterlas en las funciones de creación del vertical
1.0
Motor detección de fugas - Lanzamos una tarea en el processing que llame a una función plpgsql, aplique las reglas de fuga para cada sector y actualice la tabla de resultados de fugas. ### Reglas de fuga En una tabla tenemos la configuración de reglas de las fugas "leak_rules" . Algo como: q | p |  time | status 20 | null | 3600 |  1 40 |  -40 | 7200 | 2 status=1: Anomalía de consumo status=2: Peligro de fuga La tabla de last_data de los sectores debe tener una columna que sea "leak_status" y que tenga actualizado el último estado. Si es ok el valor debe ser 0. ### Tabla de resultados Tenemos una tabla con el histórico de leaks por sectores. "leak_sector_historic". Esta tabla se actualiza cuando una fuga cambia de estado. sector_id  | TimeInstant | status | rule Esta tabla guarda los cambios de estado de los sectores. Cuando hay un cambio de estado se crea un nuevo registro para el nuevo estado. Ejemplo: 1. Se detecta una anomalía de consumo. Se añade este registro: "1;2017-01-01T00:00Z;1;'Se detecta un consumo superior al 20%'" 2. Se detecta a la hora siguiente que la anomalía se ha corregido. Se actualiza el registro anterior "1;2017-01-01T00:00Z;1;'Se detecta un consumo superior al 20%'" Se añade un nuevo registro "1;2017-01-01T01:00Z;0;null" Hay que generar una variable Urbo para el campo status de esta tabla. La generación de tablas hay que meterlas en las funciones de creación del vertical
process
motor detección de fugas lanzamos una tarea en el processing que llame a una función plpgsql aplique las reglas de fuga para cada sector y actualice la tabla de resultados de fugas reglas de fuga en una tabla tenemos la configuración de reglas de las fugas leak rules algo como q p  time status null       status anomalía de consumo status peligro de fuga la tabla de last data de los sectores debe tener una columna que sea leak status y que tenga actualizado el último estado si es ok el valor debe ser tabla de resultados tenemos una tabla con el histórico de leaks por sectores leak sector historic esta tabla se actualiza cuando una fuga cambia de estado sector id   timeinstant status  rule esta tabla guarda los cambios de estado de los sectores cuando hay un cambio de estado se crea un nuevo registro para el nuevo estado ejemplo se detecta una anomalía de consumo se añade este registro se detecta un consumo superior al se detecta a la hora siguiente que la anomalía se ha corregido se actualiza el registro anterior se detecta un consumo superior al se añade un nuevo registro null hay que generar una variable urbo para el campo status de esta tabla la generación de tablas hay que meterlas en las funciones de creación del vertical
1
44,294
2,903,190,402
IssuesEvent
2015-06-18 12:01:17
MarcusWolschon/osmeditor4android
https://api.github.com/repos/MarcusWolschon/osmeditor4android
opened
Test against M preview
Medium Priority Other Task Work started
Test against M (marzipan?) preview. Mostly this means checking what happens when required permissions are revoked.
1.0
Test against M preview - Test against M (marzipan?) preview. Mostly this means checking what happens when required permissions are revoked.
non_process
test against m preview test against m marzipan preview mostly this means checking what happens when required permissions are revoked
0
131,778
10,709,516,940
IssuesEvent
2019-10-24 22:24:15
Azure/azure-sdk-for-python
https://api.github.com/repos/Azure/azure-sdk-for-python
closed
Logging instructions do not work
Client KeyVault blocking-release test_needed
For the key vault UX study I wanted to try and setup network trace logging with the keyvault. I used an instructions in preview.4 Readme.md to setup logging. I noticed 2 issues: 1)client = KeyClient(vault_endpoint=url, credential=credential, logging_enable=True) - constructor does not accept logging as an argument 2) when passine logging_enable to a method and follow rest of the setup instruction, logging trace does not appear on the console
1.0
Logging instructions do not work - For the key vault UX study I wanted to try and setup network trace logging with the keyvault. I used an instructions in preview.4 Readme.md to setup logging. I noticed 2 issues: 1)client = KeyClient(vault_endpoint=url, credential=credential, logging_enable=True) - constructor does not accept logging as an argument 2) when passine logging_enable to a method and follow rest of the setup instruction, logging trace does not appear on the console
non_process
logging instructions do not work for the key vault ux study i wanted to try and setup network trace logging with the keyvault i used an instructions in preview readme md to setup logging i noticed issues client keyclient vault endpoint url credential credential logging enable true constructor does not accept logging as an argument when passine logging enable to a method and follow rest of the setup instruction logging trace does not appear on the console
0
11,676
14,535,490,628
IssuesEvent
2020-12-15 05:43:50
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
feature] Processing raster calc: add missing btns and validate
3.12 Automatic new feature Processing Alg
Original commit: https://github.com/qgis/QGIS/commit/29824c1765262e821cdac447d13e06481a993a47 by web-flow [feature] Processing raster calc: add missing btns and validate
1.0
feature] Processing raster calc: add missing btns and validate - Original commit: https://github.com/qgis/QGIS/commit/29824c1765262e821cdac447d13e06481a993a47 by web-flow [feature] Processing raster calc: add missing btns and validate
process
feature processing raster calc add missing btns and validate original commit by web flow processing raster calc add missing btns and validate
1
55,516
3,073,628,980
IssuesEvent
2015-08-19 23:12:32
RobotiumTech/robotium
https://api.github.com/repos/RobotiumTech/robotium
closed
Unable to click a buttons in WebViews by using Xpath,tagName ,etc in By Except id .
bug enhancement imported invalid Priority-Medium
_From [anbarasa...@photoninfotech.net](https://code.google.com/u/110883577902740873760/) on March 25, 2013 05:54:51_ Hi Groups, in Android Hybrid Application , unable to click the buttons in web view Even we have By class in robotium 4.0 jar, only id is working fine , but Xpath should work . Example code : public void testVerificationRegister() throws TestException { soloMain.waitForActivity("MainActivity", 2000); //soloMain.getCurrentWebElements(); //soloMain.sleep(2000); By by=By.xpath("//a[contains(text(),'Browse')]"); soloMain.clickOnWebElement(by); (unable to click the application) /* soloMain.clickOnText("Login"); soloMain.sleep(2000); By FirstName = By.id("regfirstname"); soloMain.typeTextInWebElement(FirstName, "Robo"); soloMain.sleep(2000); soloMain.getCurrentWebElements(); By LastName =By.id("reglastname"); soloMain.typeTextInWebElement(LastName, "Robo"); soloMain.getCurrentWebElements(); By EmailId =By.id("regemail"); soloMain.typeTextInWebElement(EmailId, "Robo@gmail.com"); soloMain.getCurrentWebElements(); By passward =By.id("regpassword"); soloMain.typeTextInWebElement(passward, "Robo123"); soloMain.getCurrentWebElements(); By Phonenumber =By.id("regphonenumber"); soloMain.typeTextInWebElement(Phonenumber, "9943113592"); soloMain.getCurrentWebElements(); soloMain.clickOnText("Login"); By Logmail =By.id("logEmail"); soloMain.typeTextInWebElement(Logmail, "robo@gmail.com"); soloMain.getCurrentWebElements(); By pass =By.id("logpassword"); soloMain.typeTextInWebElement(pass, "robo123"); soloMain.getCurrentWebElements(); By by=By.xpath("//div[6]/div/div/div/a"); soloMain.clickOnWebElement(by); */ //soloMain.clickOnText("Submit"); //EditFields=(EditText)soloMain.getEditText(0); // soloMain.enterText(EditFields, "Anbarasu"); // soloMain.goBack(); //soloMain.clickOnText("Browse"); //soloMain.sleep(2000); //soloMain.goBack(); //soloMain.clickOnText("Register"); } _Original issue: http://code.google.com/p/robotium/issues/detail?id=424_
1.0
Unable to click a buttons in WebViews by using Xpath,tagName ,etc in By Except id . - _From [anbarasa...@photoninfotech.net](https://code.google.com/u/110883577902740873760/) on March 25, 2013 05:54:51_ Hi Groups, in Android Hybrid Application , unable to click the buttons in web view Even we have By class in robotium 4.0 jar, only id is working fine , but Xpath should work . Example code : public void testVerificationRegister() throws TestException { soloMain.waitForActivity("MainActivity", 2000); //soloMain.getCurrentWebElements(); //soloMain.sleep(2000); By by=By.xpath("//a[contains(text(),'Browse')]"); soloMain.clickOnWebElement(by); (unable to click the application) /* soloMain.clickOnText("Login"); soloMain.sleep(2000); By FirstName = By.id("regfirstname"); soloMain.typeTextInWebElement(FirstName, "Robo"); soloMain.sleep(2000); soloMain.getCurrentWebElements(); By LastName =By.id("reglastname"); soloMain.typeTextInWebElement(LastName, "Robo"); soloMain.getCurrentWebElements(); By EmailId =By.id("regemail"); soloMain.typeTextInWebElement(EmailId, "Robo@gmail.com"); soloMain.getCurrentWebElements(); By passward =By.id("regpassword"); soloMain.typeTextInWebElement(passward, "Robo123"); soloMain.getCurrentWebElements(); By Phonenumber =By.id("regphonenumber"); soloMain.typeTextInWebElement(Phonenumber, "9943113592"); soloMain.getCurrentWebElements(); soloMain.clickOnText("Login"); By Logmail =By.id("logEmail"); soloMain.typeTextInWebElement(Logmail, "robo@gmail.com"); soloMain.getCurrentWebElements(); By pass =By.id("logpassword"); soloMain.typeTextInWebElement(pass, "robo123"); soloMain.getCurrentWebElements(); By by=By.xpath("//div[6]/div/div/div/a"); soloMain.clickOnWebElement(by); */ //soloMain.clickOnText("Submit"); //EditFields=(EditText)soloMain.getEditText(0); // soloMain.enterText(EditFields, "Anbarasu"); // soloMain.goBack(); //soloMain.clickOnText("Browse"); //soloMain.sleep(2000); //soloMain.goBack(); //soloMain.clickOnText("Register"); } _Original issue: http://code.google.com/p/robotium/issues/detail?id=424_
non_process
unable to click a buttons in webviews by using xpath tagname etc in by except id from on march hi groups in android hybrid application unable to click the buttons in web view even we have by class in robotium jar only id is working fine but xpath should work example code public void testverificationregister throws testexception solomain waitforactivity mainactivity solomain getcurrentwebelements solomain sleep by by by xpath a solomain clickonwebelement by unable to click the application solomain clickontext login solomain sleep by firstname by id regfirstname solomain typetextinwebelement firstname robo solomain sleep solomain getcurrentwebelements by lastname by id reglastname solomain typetextinwebelement lastname robo solomain getcurrentwebelements by emailid by id regemail solomain typetextinwebelement emailid robo gmail com solomain getcurrentwebelements by passward by id regpassword solomain typetextinwebelement passward solomain getcurrentwebelements by phonenumber by id regphonenumber solomain typetextinwebelement phonenumber solomain getcurrentwebelements solomain clickontext login by logmail by id logemail solomain typetextinwebelement logmail robo gmail com solomain getcurrentwebelements by pass by id logpassword solomain typetextinwebelement pass solomain getcurrentwebelements by by by xpath div div div div a solomain clickonwebelement by solomain clickontext submit editfields edittext solomain getedittext solomain entertext editfields anbarasu solomain goback solomain clickontext browse solomain sleep solomain goback solomain clickontext register original issue
0
190,105
15,217,468,411
IssuesEvent
2021-02-17 16:38:50
keyboardTC/my-new-repo
https://api.github.com/repos/keyboardTC/my-new-repo
closed
Improve repository by adding documentation
documentation enhancement
The repository does not have documentation to explain what it is for. Document can help new visitor or view to understand the project upfront
1.0
Improve repository by adding documentation - The repository does not have documentation to explain what it is for. Document can help new visitor or view to understand the project upfront
non_process
improve repository by adding documentation the repository does not have documentation to explain what it is for document can help new visitor or view to understand the project upfront
0
19,824
26,212,157,649
IssuesEvent
2023-01-04 07:46:03
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing algorithms hang when accessing authManager without loaded master key
Processing Bug
### What is the bug or the crash? When a processing algorithm makes use of the authManager and the master password is not in the cache, the user is prompted for the master password. However, when entering the correct password, the algorithm then hangs. ### Steps to reproduce the issue 1. Make sure the master password can't be retrieved from the password manager automatically by unchecking `Integrate master password with you password manager` in `Settings>Authentication>Management>Utilities` 2. Make sure the master password is not set by running `QgsApplication.instance().authManager().clearMasterPassword()` in the console 3. Create a python algorithm containing the following snippet ```python from qgis.core import QgsProcessingAlgorithm, QgsApplication class ExampleProcessingAlgorithm(QgsProcessingAlgorithm): # ... def processAlgorithm(self, parameters, context, feedback): # ... manager = QgsApplication.instance().authManager() value = manager.authSetting("MY_KEY", None, True) feedback.pushInfo(f"Auth setting `MY_KEY` is {value}" ``` 4. Run the algorithm 5. When prompted, enter your master password 6. It hangs ### Versions QGIS version 3.28.1-Firenze QGIS code revision fde3b8fbb8c Qt version 5.15.3 Python version 3.9.5 GDAL/OGR version 3.6.0 PROJ version 9.1.0 EPSG Registry database version v10.074 (2022-08-01) GEOS version 3.11.1-CAPI-1.17.1 SQLite version 3.39.4 PDAL version 2.4.3 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.6 QScintilla2 version 2.13.1 OS version Windows 10 Version 2009 Active Python plugins plugin_reloader 0.9.2 QuickWKT 3.1 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.6 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context I guess it has to do with threading, since when running the exact same algorithm from python using processing.run, it does not hang.
1.0
Processing algorithms hang when accessing authManager without loaded master key - ### What is the bug or the crash? When a processing algorithm makes use of the authManager and the master password is not in the cache, the user is prompted for the master password. However, when entering the correct password, the algorithm then hangs. ### Steps to reproduce the issue 1. Make sure the master password can't be retrieved from the password manager automatically by unchecking `Integrate master password with you password manager` in `Settings>Authentication>Management>Utilities` 2. Make sure the master password is not set by running `QgsApplication.instance().authManager().clearMasterPassword()` in the console 3. Create a python algorithm containing the following snippet ```python from qgis.core import QgsProcessingAlgorithm, QgsApplication class ExampleProcessingAlgorithm(QgsProcessingAlgorithm): # ... def processAlgorithm(self, parameters, context, feedback): # ... manager = QgsApplication.instance().authManager() value = manager.authSetting("MY_KEY", None, True) feedback.pushInfo(f"Auth setting `MY_KEY` is {value}" ``` 4. Run the algorithm 5. When prompted, enter your master password 6. It hangs ### Versions QGIS version 3.28.1-Firenze QGIS code revision fde3b8fbb8c Qt version 5.15.3 Python version 3.9.5 GDAL/OGR version 3.6.0 PROJ version 9.1.0 EPSG Registry database version v10.074 (2022-08-01) GEOS version 3.11.1-CAPI-1.17.1 SQLite version 3.39.4 PDAL version 2.4.3 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.6 QScintilla2 version 2.13.1 OS version Windows 10 Version 2009 Active Python plugins plugin_reloader 0.9.2 QuickWKT 3.1 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.6 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context I guess it has to do with threading, since when running the exact same algorithm from python using processing.run, it does not hang.
process
processing algorithms hang when accessing authmanager without loaded master key what is the bug or the crash when a processing algorithm makes use of the authmanager and the master password is not in the cache the user is prompted for the master password however when entering the correct password the algorithm then hangs steps to reproduce the issue make sure the master password can t be retrieved from the password manager automatically by unchecking integrate master password with you password manager in settings authentication management utilities make sure the master password is not set by running qgsapplication instance authmanager clearmasterpassword in the console create a python algorithm containing the following snippet python from qgis core import qgsprocessingalgorithm qgsapplication class exampleprocessingalgorithm qgsprocessingalgorithm def processalgorithm self parameters context feedback manager qgsapplication instance authmanager value manager authsetting my key none true feedback pushinfo f auth setting my key is value run the algorithm when prompted enter your master password it hangs versions qgis version firenze qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version active python plugins plugin reloader quickwkt db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context i guess it has to do with threading since when running the exact same algorithm from python using processing run it does not hang
1
13,891
16,655,792,207
IssuesEvent
2021-06-05 13:55:09
laugharn/link
https://api.github.com/repos/laugharn/link
opened
Cleaner API Handling
kind/improvement process/selected size/sm team/back
Our API handlers are a little messy, we can clean them up. - [x] Use [next-connect](https://npm.im/next-connect) which has a nice, familiar pattern and works well with [next-iron-session](https://npm.im/next-iron-session) - [x] Create a default handler with session middleware for easy reuse - [x] Move route functions to lib files - [x] Update dependencies In the future we might want to investigate using a destructured single file API instead of having index, [id], etc. But we can revisit that.
1.0
Cleaner API Handling - Our API handlers are a little messy, we can clean them up. - [x] Use [next-connect](https://npm.im/next-connect) which has a nice, familiar pattern and works well with [next-iron-session](https://npm.im/next-iron-session) - [x] Create a default handler with session middleware for easy reuse - [x] Move route functions to lib files - [x] Update dependencies In the future we might want to investigate using a destructured single file API instead of having index, [id], etc. But we can revisit that.
process
cleaner api handling our api handlers are a little messy we can clean them up use which has a nice familiar pattern and works well with create a default handler with session middleware for easy reuse move route functions to lib files update dependencies in the future we might want to investigate using a destructured single file api instead of having index etc but we can revisit that
1
6,518
9,605,131,794
IssuesEvent
2019-05-10 22:25:59
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Torch.multiprocessing pool hangs in Jupyter notebooks
module: multiprocessing
## 🐛 Bug Running pool.map() hangs (Torch 1.0.0) ## To Reproduce ``` import torch import torch.nn as nn from torch.multiprocessing import Pool, set_start_method try: set_start_method('spawn') except RuntimeError: pass class Dummy: def __init__(self, device): self.device = device self.layer = nn.Linear(100, 100).to(self.device).share_memory() # removing share_memory doesn't have any effects def run(self): self.layer(torch.rand(1000, 100).to(self.device)) def run_steps(agent): for k in range(100): agent.run() n_gpus = 2 agents = [Dummy("cuda:%d" % k) for k in range(n_gpus)] # THIS HANGS pool = Pool(n_gpus) pool.map(run_steps, agents) ``` If instead I use threads, everything works: ``` # This works: from multiprocessing.dummy import Pool as dThreadPool pool = dThreadPool(n_gpus) pool.map(run_steps, agents) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior I expect the code to run without hanging ## Environment Pytorch 1.0.0 OS: Ubuntu 16.04.3 LTS GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.11) 5.4.0 20160609 Installed via pip Python version: 3.5 CUDA 9.0 GPU models and configuration: GPU 0: Tesla P100-PCIE-16GB GPU 1: Tesla P100-PCIE-16GB Nvidia driver version: 384.145 cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.3.1 ## Additional context This happens when running in a jupyter notebook Might be related to #16954
1.0
Torch.multiprocessing pool hangs in Jupyter notebooks - ## 🐛 Bug Running pool.map() hangs (Torch 1.0.0) ## To Reproduce ``` import torch import torch.nn as nn from torch.multiprocessing import Pool, set_start_method try: set_start_method('spawn') except RuntimeError: pass class Dummy: def __init__(self, device): self.device = device self.layer = nn.Linear(100, 100).to(self.device).share_memory() # removing share_memory doesn't have any effects def run(self): self.layer(torch.rand(1000, 100).to(self.device)) def run_steps(agent): for k in range(100): agent.run() n_gpus = 2 agents = [Dummy("cuda:%d" % k) for k in range(n_gpus)] # THIS HANGS pool = Pool(n_gpus) pool.map(run_steps, agents) ``` If instead I use threads, everything works: ``` # This works: from multiprocessing.dummy import Pool as dThreadPool pool = dThreadPool(n_gpus) pool.map(run_steps, agents) ``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior I expect the code to run without hanging ## Environment Pytorch 1.0.0 OS: Ubuntu 16.04.3 LTS GCC version: (Ubuntu 5.4.0-6ubuntu1~16.04.11) 5.4.0 20160609 Installed via pip Python version: 3.5 CUDA 9.0 GPU models and configuration: GPU 0: Tesla P100-PCIE-16GB GPU 1: Tesla P100-PCIE-16GB Nvidia driver version: 384.145 cuDNN version: /usr/lib/x86_64-linux-gnu/libcudnn.so.7.3.1 ## Additional context This happens when running in a jupyter notebook Might be related to #16954
process
torch multiprocessing pool hangs in jupyter notebooks 🐛 bug running pool map hangs torch to reproduce import torch import torch nn as nn from torch multiprocessing import pool set start method try set start method spawn except runtimeerror pass class dummy def init self device self device device self layer nn linear to self device share memory removing share memory doesn t have any effects def run self self layer torch rand to self device def run steps agent for k in range agent run n gpus agents this hangs pool pool n gpus pool map run steps agents if instead i use threads everything works this works from multiprocessing dummy import pool as dthreadpool pool dthreadpool n gpus pool map run steps agents expected behavior i expect the code to run without hanging environment pytorch os ubuntu lts gcc version ubuntu installed via pip python version cuda gpu models and configuration gpu tesla pcie gpu tesla pcie nvidia driver version cudnn version usr lib linux gnu libcudnn so additional context this happens when running in a jupyter notebook might be related to
1
8,784
11,902,303,645
IssuesEvent
2020-03-30 13:47:02
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Failing test-stdio-readable-writable test on Windows
CI / flaky test child_process test windows
* **Version**: v14.0.0-pre * **Platform**: Microsoft Windows NT 10.0.18362.0 x64 * **Subsystem**: child_process * **Env**: Visual Studio 2019 with C++ support, PowershellCore on Windows Terminal (Preview), nasm and python 2.7 installed via chocolatey ``` === release test-stdio-readable-writable === Path: parallel/test-stdio-readable-writable Command: C:\Users\saito\code\node\out\Release\node.exe C:\Users\saito\code\node\test\parallel\test-stdio-readable-writable.js --- TIMEOUT --- ```
1.0
Failing test-stdio-readable-writable test on Windows - * **Version**: v14.0.0-pre * **Platform**: Microsoft Windows NT 10.0.18362.0 x64 * **Subsystem**: child_process * **Env**: Visual Studio 2019 with C++ support, PowershellCore on Windows Terminal (Preview), nasm and python 2.7 installed via chocolatey ``` === release test-stdio-readable-writable === Path: parallel/test-stdio-readable-writable Command: C:\Users\saito\code\node\out\Release\node.exe C:\Users\saito\code\node\test\parallel\test-stdio-readable-writable.js --- TIMEOUT --- ```
process
failing test stdio readable writable test on windows version pre platform microsoft windows nt subsystem child process env visual studio with c support powershellcore on windows terminal preview nasm and python installed via chocolatey release test stdio readable writable path parallel test stdio readable writable command c users saito code node out release node exe c users saito code node test parallel test stdio readable writable js timeout
1
9,483
12,477,849,073
IssuesEvent
2020-05-29 15:38:00
MHRA/products
https://api.github.com/repos/MHRA/products
closed
Field values should be converted to ALL CAPS by API
EPIC - PARs process
Allow capital letters and non-capital letters in fields but all results should be capitalised when processed See point 2 in https://github.com/MHRA/products/issues/845
1.0
Field values should be converted to ALL CAPS by API - Allow capital letters and non-capital letters in fields but all results should be capitalised when processed See point 2 in https://github.com/MHRA/products/issues/845
process
field values should be converted to all caps by api allow capital letters and non capital letters in fields but all results should be capitalised when processed see point in
1
444,949
12,823,866,336
IssuesEvent
2020-07-06 12:32:01
CDH-Studio/UpSkill
https://api.github.com/repos/CDH-Studio/UpSkill
closed
Sort and display Developer goals and School dropdowns in alphabetical order
Low Priority enhancement
**Describe the solution you'd like** During profile setup or edit, when users search for developmental goals (in the Personal Growth Interests section) or schools (in the Qualification section), the dropdown options should be displayed alphabetically (ascending). **Additional context** Screenshot showing non-alphabetical ordering of developmental goals: <img width="978" alt="Screen Shot 2020-05-07 at 11 28 35 AM" src="https://user-images.githubusercontent.com/43188335/81313459-e5307700-9055-11ea-9005-1af1e99309ad.png"> Screenshot showing non-alphabetical ordering of schools: <img width="965" alt="Screen Shot 2020-05-07 at 11 29 11 AM" src="https://user-images.githubusercontent.com/43188335/81313535-faa5a100-9055-11ea-9739-946be20c8729.png">
1.0
Sort and display Developer goals and School dropdowns in alphabetical order - **Describe the solution you'd like** During profile setup or edit, when users search for developmental goals (in the Personal Growth Interests section) or schools (in the Qualification section), the dropdown options should be displayed alphabetically (ascending). **Additional context** Screenshot showing non-alphabetical ordering of developmental goals: <img width="978" alt="Screen Shot 2020-05-07 at 11 28 35 AM" src="https://user-images.githubusercontent.com/43188335/81313459-e5307700-9055-11ea-9005-1af1e99309ad.png"> Screenshot showing non-alphabetical ordering of schools: <img width="965" alt="Screen Shot 2020-05-07 at 11 29 11 AM" src="https://user-images.githubusercontent.com/43188335/81313535-faa5a100-9055-11ea-9739-946be20c8729.png">
non_process
sort and display developer goals and school dropdowns in alphabetical order describe the solution you d like during profile setup or edit when users search for developmental goals in the personal growth interests section or schools in the qualification section the dropdown options should be displayed alphabetically ascending additional context screenshot showing non alphabetical ordering of developmental goals img width alt screen shot at am src screenshot showing non alphabetical ordering of schools img width alt screen shot at am src
0
2,085
4,912,630,948
IssuesEvent
2016-11-23 09:46:18
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
Dynamic table is not greyed out in completed form
browser: all bug comp: activiti-processList
Dynamic table not greyed out in a completed form attached to a start event <img width="656" alt="screen shot 2016-11-23 at 09 44 29" src="https://cloud.githubusercontent.com/assets/13200338/20556996/83e276b4-b161-11e6-9333-70903601a310.png">
1.0
Dynamic table is not greyed out in completed form - Dynamic table not greyed out in a completed form attached to a start event <img width="656" alt="screen shot 2016-11-23 at 09 44 29" src="https://cloud.githubusercontent.com/assets/13200338/20556996/83e276b4-b161-11e6-9333-70903601a310.png">
process
dynamic table is not greyed out in completed form dynamic table not greyed out in a completed form attached to a start event img width alt screen shot at src
1
9,932
12,969,908,811
IssuesEvent
2020-07-21 08:31:17
keep-network/keep-core
https://api.github.com/repos/keep-network/keep-core
closed
Allow bridging already-staked tokens on new staking contract
process & client team ⛓chain
This is a light proposal for how to allow existing stakers to port their stake over to the updated staking contract we'll be supporting. ## Background We'll be leveraging the tBTC deposit pause to deploy a new staking contract with a few features that weren't possible in the v1 release of the Keep core contracts. In particular, this staking contract will support “top ups”, meaning the ability to delegate additional tokens to an existing delegation relationship (i.e., an existing operator/beneficiary/authorizer triplet). This introduces a key complication, however: tokens already delegated through the current staking contract must remain locked in that contract for 60 days, the length of the undelegation period. This means that existing token owners who have used the old staking contract must wait 60 days to port those balances over to the new contract. This issue outlines a proposal for how to deal with this by providing some additional liquidity from the primary token supply. ## Proposal The core goal is for the Keep team to make an amount of KEEP tokens available to temporarily stake on behalf of a token owner, provided that they have those same tokens currently locked up in the old staking contract. Undelegation of all tokens in the new staking contract would be blocked until the temporarily-staked tokens provided by the Keep team are repaid by the token owner. ### Mechanism A number of KEEP tokens equivalent to the currently-staked token amount will be locked in a special contract referred to here as the `StakingPortBacker` contract. #### “Copying” stake from the old contract to the new contract. The new staking contract will gain an additional function, `copyStake(address operator)`. `copyStake(address operator)` will request the copy from the `StakingPortBacker` contract, which will look up the given operator's information on the old staking contract, particularly its balance and owner. Only the owner of the relationship can copy it to the new staking contract, the relationship can only be copied once, and undelegation must not be complete on that relationship. The copying operation will transfer an amount of tokens equal to the old staking contract's delegation amount for this relationship from `StakingPortBacker`'s balance to the new staking contract. Once this operation is complete, the new staking contract will configure that relationship on its end to mirror the old one and set a flag on the relationship, referred to here as `hasBackedBalance`. Copying will not be allowed if the token owner has one or more existing delegations on the new staking contract, unless there is exactly one existing delegation to the same operator/authorizer/beneficiary triplet. #### Limitations on recovering stake and new delegations The new staking contract's `recoverStake` operation will prevent recovery of undelegated stake if the `hasBackedBalance` boolean on the relationship whose stake is being recovered is `true`. That is, the `StakingPortBacker` must be repaid (see below) before any stake can be recovered for that relationship. New delegations will not be allowed for a token owner with a `hasBackedBalance` delegation. Top ups (adding new tokens to the existing delegation relationship) will be allowed. #### Repaying backed balances The `StakingPortBacker` will have a function available, `repay(address operator)`, which repays the backed balance for a given operator relationship. It will call a function on the new token staking contract, `unmarkBackedBalance(address operator)`, which only `StakingPortBacker` is authorized to call. This will clear the `hasBackedBalance` flag on the relationship and re-enable both new delegation relationships and stake recovery for that relationship. #### Time limit To encourage transition over a reasonable time frame, the `StakingPortBacker` will only allow stake copying for a limited period of time, after which copying will no longer be permitted. It will at all times provide a function, `withdraw`, that will allow the contract owner to withdraw tokens from the balance the contract has available to back stake copying. New tokens can always be transferred to the contract via a standard `transfer` call. #### Expected usage The expected mode of usage for users who have already staked on the old contract is to: - Copy stake. - Undelegate from the old staking contract. - Operate as normal on the new staking contract, possibly adding more token balance over time. - Once the old staking contract's undelegated balance is available for recovery, recover it and use the recovered tokens to repay the `StakingPortBacker` contract. ### Limitations The greatest potential pitfall here is the already-large size of the staking contracts. Adding this functionality could bump us against contract size limits, something we've struggled with in a few other places as well. The above proposal tries to keep as much code out of the staking contract as possible, but there is still some complexity to be added in several places.
1.0
Allow bridging already-staked tokens on new staking contract - This is a light proposal for how to allow existing stakers to port their stake over to the updated staking contract we'll be supporting. ## Background We'll be leveraging the tBTC deposit pause to deploy a new staking contract with a few features that weren't possible in the v1 release of the Keep core contracts. In particular, this staking contract will support “top ups”, meaning the ability to delegate additional tokens to an existing delegation relationship (i.e., an existing operator/beneficiary/authorizer triplet). This introduces a key complication, however: tokens already delegated through the current staking contract must remain locked in that contract for 60 days, the length of the undelegation period. This means that existing token owners who have used the old staking contract must wait 60 days to port those balances over to the new contract. This issue outlines a proposal for how to deal with this by providing some additional liquidity from the primary token supply. ## Proposal The core goal is for the Keep team to make an amount of KEEP tokens available to temporarily stake on behalf of a token owner, provided that they have those same tokens currently locked up in the old staking contract. Undelegation of all tokens in the new staking contract would be blocked until the temporarily-staked tokens provided by the Keep team are repaid by the token owner. ### Mechanism A number of KEEP tokens equivalent to the currently-staked token amount will be locked in a special contract referred to here as the `StakingPortBacker` contract. #### “Copying” stake from the old contract to the new contract. The new staking contract will gain an additional function, `copyStake(address operator)`. `copyStake(address operator)` will request the copy from the `StakingPortBacker` contract, which will look up the given operator's information on the old staking contract, particularly its balance and owner. Only the owner of the relationship can copy it to the new staking contract, the relationship can only be copied once, and undelegation must not be complete on that relationship. The copying operation will transfer an amount of tokens equal to the old staking contract's delegation amount for this relationship from `StakingPortBacker`'s balance to the new staking contract. Once this operation is complete, the new staking contract will configure that relationship on its end to mirror the old one and set a flag on the relationship, referred to here as `hasBackedBalance`. Copying will not be allowed if the token owner has one or more existing delegations on the new staking contract, unless there is exactly one existing delegation to the same operator/authorizer/beneficiary triplet. #### Limitations on recovering stake and new delegations The new staking contract's `recoverStake` operation will prevent recovery of undelegated stake if the `hasBackedBalance` boolean on the relationship whose stake is being recovered is `true`. That is, the `StakingPortBacker` must be repaid (see below) before any stake can be recovered for that relationship. New delegations will not be allowed for a token owner with a `hasBackedBalance` delegation. Top ups (adding new tokens to the existing delegation relationship) will be allowed. #### Repaying backed balances The `StakingPortBacker` will have a function available, `repay(address operator)`, which repays the backed balance for a given operator relationship. It will call a function on the new token staking contract, `unmarkBackedBalance(address operator)`, which only `StakingPortBacker` is authorized to call. This will clear the `hasBackedBalance` flag on the relationship and re-enable both new delegation relationships and stake recovery for that relationship. #### Time limit To encourage transition over a reasonable time frame, the `StakingPortBacker` will only allow stake copying for a limited period of time, after which copying will no longer be permitted. It will at all times provide a function, `withdraw`, that will allow the contract owner to withdraw tokens from the balance the contract has available to back stake copying. New tokens can always be transferred to the contract via a standard `transfer` call. #### Expected usage The expected mode of usage for users who have already staked on the old contract is to: - Copy stake. - Undelegate from the old staking contract. - Operate as normal on the new staking contract, possibly adding more token balance over time. - Once the old staking contract's undelegated balance is available for recovery, recover it and use the recovered tokens to repay the `StakingPortBacker` contract. ### Limitations The greatest potential pitfall here is the already-large size of the staking contracts. Adding this functionality could bump us against contract size limits, something we've struggled with in a few other places as well. The above proposal tries to keep as much code out of the staking contract as possible, but there is still some complexity to be added in several places.
process
allow bridging already staked tokens on new staking contract this is a light proposal for how to allow existing stakers to port their stake over to the updated staking contract we ll be supporting background we ll be leveraging the tbtc deposit pause to deploy a new staking contract with a few features that weren t possible in the release of the keep core contracts in particular this staking contract will support “top ups” meaning the ability to delegate additional tokens to an existing delegation relationship i e an existing operator beneficiary authorizer triplet this introduces a key complication however tokens already delegated through the current staking contract must remain locked in that contract for days the length of the undelegation period this means that existing token owners who have used the old staking contract must wait days to port those balances over to the new contract this issue outlines a proposal for how to deal with this by providing some additional liquidity from the primary token supply proposal the core goal is for the keep team to make an amount of keep tokens available to temporarily stake on behalf of a token owner provided that they have those same tokens currently locked up in the old staking contract undelegation of all tokens in the new staking contract would be blocked until the temporarily staked tokens provided by the keep team are repaid by the token owner mechanism a number of keep tokens equivalent to the currently staked token amount will be locked in a special contract referred to here as the stakingportbacker contract “copying” stake from the old contract to the new contract the new staking contract will gain an additional function copystake address operator copystake address operator will request the copy from the stakingportbacker contract which will look up the given operator s information on the old staking contract particularly its balance and owner only the owner of the relationship can copy it to the new staking contract the relationship can only be copied once and undelegation must not be complete on that relationship the copying operation will transfer an amount of tokens equal to the old staking contract s delegation amount for this relationship from stakingportbacker s balance to the new staking contract once this operation is complete the new staking contract will configure that relationship on its end to mirror the old one and set a flag on the relationship referred to here as hasbackedbalance copying will not be allowed if the token owner has one or more existing delegations on the new staking contract unless there is exactly one existing delegation to the same operator authorizer beneficiary triplet limitations on recovering stake and new delegations the new staking contract s recoverstake operation will prevent recovery of undelegated stake if the hasbackedbalance boolean on the relationship whose stake is being recovered is true that is the stakingportbacker must be repaid see below before any stake can be recovered for that relationship new delegations will not be allowed for a token owner with a hasbackedbalance delegation top ups adding new tokens to the existing delegation relationship will be allowed repaying backed balances the stakingportbacker will have a function available repay address operator which repays the backed balance for a given operator relationship it will call a function on the new token staking contract unmarkbackedbalance address operator which only stakingportbacker is authorized to call this will clear the hasbackedbalance flag on the relationship and re enable both new delegation relationships and stake recovery for that relationship time limit to encourage transition over a reasonable time frame the stakingportbacker will only allow stake copying for a limited period of time after which copying will no longer be permitted it will at all times provide a function withdraw that will allow the contract owner to withdraw tokens from the balance the contract has available to back stake copying new tokens can always be transferred to the contract via a standard transfer call expected usage the expected mode of usage for users who have already staked on the old contract is to copy stake undelegate from the old staking contract operate as normal on the new staking contract possibly adding more token balance over time once the old staking contract s undelegated balance is available for recovery recover it and use the recovered tokens to repay the stakingportbacker contract limitations the greatest potential pitfall here is the already large size of the staking contracts adding this functionality could bump us against contract size limits something we ve struggled with in a few other places as well the above proposal tries to keep as much code out of the staking contract as possible but there is still some complexity to be added in several places
1
17,175
22,747,464,989
IssuesEvent
2022-07-07 10:24:44
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
opened
Introduce ProcessingScheduleService
kind/toil team/distributed team/process-automation
**Description** Part of #9600 Create an abstraction around the processing actor, since we want to remove the dependency and knowledge about the current execution model from the engine. This will simplify the tests as we can use a simple implementation for the execution and it should also help us to make further changes to the actor scheduler etc. **Todo:** - [ ] Create a new interface for the ProcessingScheduleService (with narrowed scope) - [ ] Possibily only two methods, runDelayed and runComplete take a look at the POC #9602 - [ ] Ideally we use real futures on the runComplete, but might be something we refactor later - [ ] Implement the interface, and provide certain guarantees, [take a look at POC for more details ](https://github.com/camunda/zeebe/issues/9602#issuecomment-1172029736) - [ ] Guarantee 1: We are executed after processing and committing transaction (e.g. if scheduling happened during processing) - [ ] Guarantee 2: We are not executed in parallel with processing (discussable, depends on the Writers and Result implementation) - [ ] ...
1.0
Introduce ProcessingScheduleService - **Description** Part of #9600 Create an abstraction around the processing actor, since we want to remove the dependency and knowledge about the current execution model from the engine. This will simplify the tests as we can use a simple implementation for the execution and it should also help us to make further changes to the actor scheduler etc. **Todo:** - [ ] Create a new interface for the ProcessingScheduleService (with narrowed scope) - [ ] Possibily only two methods, runDelayed and runComplete take a look at the POC #9602 - [ ] Ideally we use real futures on the runComplete, but might be something we refactor later - [ ] Implement the interface, and provide certain guarantees, [take a look at POC for more details ](https://github.com/camunda/zeebe/issues/9602#issuecomment-1172029736) - [ ] Guarantee 1: We are executed after processing and committing transaction (e.g. if scheduling happened during processing) - [ ] Guarantee 2: We are not executed in parallel with processing (discussable, depends on the Writers and Result implementation) - [ ] ...
process
introduce processingscheduleservice description part of create an abstraction around the processing actor since we want to remove the dependency and knowledge about the current execution model from the engine this will simplify the tests as we can use a simple implementation for the execution and it should also help us to make further changes to the actor scheduler etc todo create a new interface for the processingscheduleservice with narrowed scope possibily only two methods rundelayed and runcomplete take a look at the poc ideally we use real futures on the runcomplete but might be something we refactor later implement the interface and provide certain guarantees guarantee we are executed after processing and committing transaction e g if scheduling happened during processing guarantee we are not executed in parallel with processing discussable depends on the writers and result implementation
1
15,102
18,840,570,161
IssuesEvent
2021-11-11 09:04:41
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
We should be able to handle a panic (trigger error reporting) if it comes from `getDMMF`
process/candidate kind/improvement topic: cli topic: error topic: dmmf topic: cli-validate tech/engines topic: error reporting team/client
This was only confirmed using the Node-API engine when using getDMMF (to be confirmed if binary engine is also affected) As of today a panic happening in `getDmmfNodeAPI` looks like this (same ~ output for `validate`, `migrate dev` or `db push`) ``` npx prisma@latest db push Prisma schema loaded from schema.prisma Datasource "db": PostgreSQL database "repropanicvalidate", schema "public" at "localhost:5432" PostgreSQL database repropanicvalidate created at localhost:5432 Error: db error: ERROR: syntax error at or near ")" 0: sql_migration_connector::sql_database_step_applier::apply_migration at migration-engine/connectors/sql-migration-connector/src/sql_database_step_applier.rs:11 1: migration_core::api::SchemaPush at migration-engine/core/src/api.rs:184 ``` If we fix it it will look like this (here DEBUG is enabled with extra logging) <img width="671" alt="Screen Shot 2021-11-11 at 09 51 03" src="https://user-images.githubusercontent.com/1328733/141268204-493438b4-c712-4dd6-8e48-c7c2f41812bb.png"> I created a small PR to demonstrate what we need to do: https://github.com/prisma/prisma/pull/10205/files How to reproduce? See https://github.com/prisma/prisma/issues/10118#issuecomment-965641590 In a nutshell - create this schema (do not auto format) ```prisma datasource db { provider = "postgresql" url = "postgres://prisma:prisma@localhost:5432/repropanicvalidate" } generator client { provider = "prisma-client-js" } model User { id String @id @default(cuid()) @db.Char(30) referralId String @unique @db.Char(30) referral Referral? @relation("UserToReferral", fields: [referralId], references: [id]) } model Referral { id String @id @default(cuid()) @db.Char(30) user User @relation("UserToReferral") } ``` - run `prisma validate`
1.0
We should be able to handle a panic (trigger error reporting) if it comes from `getDMMF` - This was only confirmed using the Node-API engine when using getDMMF (to be confirmed if binary engine is also affected) As of today a panic happening in `getDmmfNodeAPI` looks like this (same ~ output for `validate`, `migrate dev` or `db push`) ``` npx prisma@latest db push Prisma schema loaded from schema.prisma Datasource "db": PostgreSQL database "repropanicvalidate", schema "public" at "localhost:5432" PostgreSQL database repropanicvalidate created at localhost:5432 Error: db error: ERROR: syntax error at or near ")" 0: sql_migration_connector::sql_database_step_applier::apply_migration at migration-engine/connectors/sql-migration-connector/src/sql_database_step_applier.rs:11 1: migration_core::api::SchemaPush at migration-engine/core/src/api.rs:184 ``` If we fix it it will look like this (here DEBUG is enabled with extra logging) <img width="671" alt="Screen Shot 2021-11-11 at 09 51 03" src="https://user-images.githubusercontent.com/1328733/141268204-493438b4-c712-4dd6-8e48-c7c2f41812bb.png"> I created a small PR to demonstrate what we need to do: https://github.com/prisma/prisma/pull/10205/files How to reproduce? See https://github.com/prisma/prisma/issues/10118#issuecomment-965641590 In a nutshell - create this schema (do not auto format) ```prisma datasource db { provider = "postgresql" url = "postgres://prisma:prisma@localhost:5432/repropanicvalidate" } generator client { provider = "prisma-client-js" } model User { id String @id @default(cuid()) @db.Char(30) referralId String @unique @db.Char(30) referral Referral? @relation("UserToReferral", fields: [referralId], references: [id]) } model Referral { id String @id @default(cuid()) @db.Char(30) user User @relation("UserToReferral") } ``` - run `prisma validate`
process
we should be able to handle a panic trigger error reporting if it comes from getdmmf this was only confirmed using the node api engine when using getdmmf to be confirmed if binary engine is also affected as of today a panic happening in getdmmfnodeapi looks like this same output for validate migrate dev or db push npx prisma latest db push prisma schema loaded from schema prisma datasource db postgresql database repropanicvalidate schema public at localhost postgresql database repropanicvalidate created at localhost error db error error syntax error at or near sql migration connector sql database step applier apply migration at migration engine connectors sql migration connector src sql database step applier rs migration core api schemapush at migration engine core src api rs if we fix it it will look like this here debug is enabled with extra logging img width alt screen shot at src i created a small pr to demonstrate what we need to do how to reproduce see in a nutshell create this schema do not auto format prisma datasource db provider postgresql url postgres prisma prisma localhost repropanicvalidate generator client provider prisma client js model user id string id default cuid db char referralid string unique db char referral referral relation usertoreferral fields references model referral id string id default cuid db char user user relation usertoreferral run prisma validate
1
58,157
14,242,209,092
IssuesEvent
2020-11-19 01:13:28
RG4421/react-sizeme
https://api.github.com/repos/RG4421/react-sizeme
closed
WS-2019-0019 (Medium) detected in braces-1.8.5.tgz - autoclosed
security vulnerability
## WS-2019-0019 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary> <p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p> <p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p> <p>Path to dependency file: react-sizeme/package.json</p> <p>Path to vulnerable library: react-sizeme/node_modules/braces/package.json</p> <p> Dependency Hierarchy: - jest-21.1.0.tgz (Root Library) - jest-cli-21.1.0.tgz - micromatch-2.3.11.tgz - :x: **braces-1.8.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/RG4421/react-sizeme/commit/cc2a5e77b7c2a93b4e3a04102bd6cdbfcdc4fdce">cc2a5e77b7c2a93b4e3a04102bd6cdbfcdc4fdce</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service. <p>Publish Date: 2018-02-16 <p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p> <p>Release Date: 2019-02-21</p> <p>Fix Resolution: 2.3.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"braces","packageVersion":"1.8.5","isTransitiveDependency":true,"dependencyTree":"jest:21.1.0;jest-cli:21.1.0;micromatch:2.3.11;braces:1.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.1"}],"vulnerabilityIdentifier":"WS-2019-0019","vulnerabilityDetails":"Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.","vulnerabilityUrl":"https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
True
WS-2019-0019 (Medium) detected in braces-1.8.5.tgz - autoclosed - ## WS-2019-0019 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary> <p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p> <p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p> <p>Path to dependency file: react-sizeme/package.json</p> <p>Path to vulnerable library: react-sizeme/node_modules/braces/package.json</p> <p> Dependency Hierarchy: - jest-21.1.0.tgz (Root Library) - jest-cli-21.1.0.tgz - micromatch-2.3.11.tgz - :x: **braces-1.8.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/RG4421/react-sizeme/commit/cc2a5e77b7c2a93b4e3a04102bd6cdbfcdc4fdce">cc2a5e77b7c2a93b4e3a04102bd6cdbfcdc4fdce</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service. <p>Publish Date: 2018-02-16 <p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p> <p>Release Date: 2019-02-21</p> <p>Fix Resolution: 2.3.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"braces","packageVersion":"1.8.5","isTransitiveDependency":true,"dependencyTree":"jest:21.1.0;jest-cli:21.1.0;micromatch:2.3.11;braces:1.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.1"}],"vulnerabilityIdentifier":"WS-2019-0019","vulnerabilityDetails":"Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.","vulnerabilityUrl":"https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
non_process
ws medium detected in braces tgz autoclosed ws medium severity vulnerability vulnerable library braces tgz fastest brace expansion for node js with the most complete support for the bash braces specification library home page a href path to dependency file react sizeme package json path to vulnerable library react sizeme node modules braces package json dependency hierarchy jest tgz root library jest cli tgz micromatch tgz x braces tgz vulnerable library found in head commit a href vulnerability details version of braces prior to are vulnerable to regular expression denial of service redos untrusted input may cause catastrophic backtracking while matching regular expressions this can cause the application to be unresponsive leading to denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails version of braces prior to are vulnerable to regular expression denial of service redos untrusted input may cause catastrophic backtracking while matching regular expressions this can cause the application to be unresponsive leading to denial of service vulnerabilityurl
0
26,838
20,766,080,121
IssuesEvent
2022-03-15 20:45:34
opendp/opendp
https://api.github.com/repos/opendp/opendp
closed
Merge `opendp` and `opendp-ffi` crates
OpenDP Core Effort 2 - Medium :cookie: CATEGORY: Infrastructure
We want to automatically generate FFI bindings. One of the necessary steps is to merge the `opendp` and `opendp-ffi` crates. 1. If we generate extern functions via proc-macros, the extern functions live in the opendp crate. Unfortunately, extern functions cannot reliably be re-exported from dependencies on all architectures. 2. We could try parsing a foreign crate's sources (opendp) from the opendp-ffi build.rs and generate extern functions to the `OUT_DIR`, but crate sources are somewhat isolated. It might be possible to unpack the cargo cache's .crate files for the opendp lib. This seems fragile. If we merge the crates, and put FFI behind a flag, then the build.rs could read the crate's own sources, and generate extern fn's that don't need to be re-exported.
1.0
Merge `opendp` and `opendp-ffi` crates - We want to automatically generate FFI bindings. One of the necessary steps is to merge the `opendp` and `opendp-ffi` crates. 1. If we generate extern functions via proc-macros, the extern functions live in the opendp crate. Unfortunately, extern functions cannot reliably be re-exported from dependencies on all architectures. 2. We could try parsing a foreign crate's sources (opendp) from the opendp-ffi build.rs and generate extern functions to the `OUT_DIR`, but crate sources are somewhat isolated. It might be possible to unpack the cargo cache's .crate files for the opendp lib. This seems fragile. If we merge the crates, and put FFI behind a flag, then the build.rs could read the crate's own sources, and generate extern fn's that don't need to be re-exported.
non_process
merge opendp and opendp ffi crates we want to automatically generate ffi bindings one of the necessary steps is to merge the opendp and opendp ffi crates if we generate extern functions via proc macros the extern functions live in the opendp crate unfortunately extern functions cannot reliably be re exported from dependencies on all architectures we could try parsing a foreign crate s sources opendp from the opendp ffi build rs and generate extern functions to the out dir but crate sources are somewhat isolated it might be possible to unpack the cargo cache s crate files for the opendp lib this seems fragile if we merge the crates and put ffi behind a flag then the build rs could read the crate s own sources and generate extern fn s that don t need to be re exported
0
8,209
11,404,202,509
IssuesEvent
2020-01-31 09:16:59
prisma/migrate
https://api.github.com/repos/prisma/migrate
closed
SQLite create DB fails
bug/2-confirmed kind/regression process/next-milestone
Again, with preview 19 Once I fix the schema from https://github.com/prisma/prisma2/issues/1294 Running lift up fails, 1. I can see an `undefined` in the name ![image](https://user-images.githubusercontent.com/746482/71996459-893b1a80-323c-11ea-9bf9-f420ff80800c.png) 2. Gets stuck here with no error ![image](https://user-images.githubusercontent.com/746482/71996485-96f0a000-323c-11ea-8d86-5e014349f2bb.png)
1.0
SQLite create DB fails - Again, with preview 19 Once I fix the schema from https://github.com/prisma/prisma2/issues/1294 Running lift up fails, 1. I can see an `undefined` in the name ![image](https://user-images.githubusercontent.com/746482/71996459-893b1a80-323c-11ea-9bf9-f420ff80800c.png) 2. Gets stuck here with no error ![image](https://user-images.githubusercontent.com/746482/71996485-96f0a000-323c-11ea-8d86-5e014349f2bb.png)
process
sqlite create db fails again with preview once i fix the schema from running lift up fails i can see an undefined in the name gets stuck here with no error
1
758,460
26,556,222,443
IssuesEvent
2023-01-20 12:16:04
eclipse/openvsx
https://api.github.com/repos/eclipse/openvsx
closed
Add page that lists all extensions of one author
webui server priority:medium
Microsoft marketplace has a link that shows all extensions of one publisher/author: ![demo](https://user-images.githubusercontent.com/9638156/108116571-d5e25d80-70ac-11eb-92db-12ab5bc071fa.gif) This is not a dashboard for publishers to manage or look at statistics, but for users to see if author has any other good extension(s).
1.0
Add page that lists all extensions of one author - Microsoft marketplace has a link that shows all extensions of one publisher/author: ![demo](https://user-images.githubusercontent.com/9638156/108116571-d5e25d80-70ac-11eb-92db-12ab5bc071fa.gif) This is not a dashboard for publishers to manage or look at statistics, but for users to see if author has any other good extension(s).
non_process
add page that lists all extensions of one author microsoft marketplace has a link that shows all extensions of one publisher author this is not a dashboard for publishers to manage or look at statistics but for users to see if author has any other good extension s
0
322,493
27,611,708,697
IssuesEvent
2023-03-09 16:24:26
neondatabase/neon
https://api.github.com/repos/neondatabase/neon
opened
Spurious `cargo test` failures
a/test/flaky
We've had a few slack threads on these already: - https://neondb.slack.com/archives/C04DGM6SMTM/p1673017754801879 - https://neondb.slack.com/archives/C033RQ5SPDH/p1676811125248989 They are essentially about these failures [build-neon job](https://github.com/neondatabase/neon/actions/runs/4376092920/jobs/7657714851): ``` failures: v14::xlog_utils::tests::test_find_end_of_wal_crossing_segment_followed_by_small_one v14::xlog_utils::tests::test_find_end_of_wal_last_crossing_segment v14::xlog_utils::tests::test_find_end_of_wal_simple v15::xlog_utils::tests::test_find_end_of_wal_crossing_segment_followed_by_small_one v15::xlog_utils::tests::test_find_end_of_wal_last_crossing_segment v15::xlog_utils::tests::test_find_end_of_wal_simple test result: FAILED. 25 passed; 6 failed; 0 ignored; 0 measured; 0 filtered out; finished in 2.16s error: test failed, to rerun pass `-p postgres_ffi --lib` ``` These seem quite rare.
1.0
Spurious `cargo test` failures - We've had a few slack threads on these already: - https://neondb.slack.com/archives/C04DGM6SMTM/p1673017754801879 - https://neondb.slack.com/archives/C033RQ5SPDH/p1676811125248989 They are essentially about these failures [build-neon job](https://github.com/neondatabase/neon/actions/runs/4376092920/jobs/7657714851): ``` failures: v14::xlog_utils::tests::test_find_end_of_wal_crossing_segment_followed_by_small_one v14::xlog_utils::tests::test_find_end_of_wal_last_crossing_segment v14::xlog_utils::tests::test_find_end_of_wal_simple v15::xlog_utils::tests::test_find_end_of_wal_crossing_segment_followed_by_small_one v15::xlog_utils::tests::test_find_end_of_wal_last_crossing_segment v15::xlog_utils::tests::test_find_end_of_wal_simple test result: FAILED. 25 passed; 6 failed; 0 ignored; 0 measured; 0 filtered out; finished in 2.16s error: test failed, to rerun pass `-p postgres_ffi --lib` ``` These seem quite rare.
non_process
spurious cargo test failures we ve had a few slack threads on these already they are essentially about these failures failures xlog utils tests test find end of wal crossing segment followed by small one xlog utils tests test find end of wal last crossing segment xlog utils tests test find end of wal simple xlog utils tests test find end of wal crossing segment followed by small one xlog utils tests test find end of wal last crossing segment xlog utils tests test find end of wal simple test result failed passed failed ignored measured filtered out finished in error test failed to rerun pass p postgres ffi lib these seem quite rare
0
21,062
28,010,811,148
IssuesEvent
2023-03-27 18:28:59
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Review Ocient Driver
Database/ .Epic .Team/QueryProcessor :hammer_and_wrench:
The Ocient team developed [their driver](https://github.com/Xeograph/metabase-ocient-driver) and are applying for the partnership program. To add the driver to our Cloud offerings, we need to: - [x] Code review looking for bad practices and security vulnerabilities - [x] Give them a direction to solve the failing tests - [x] As soon as we are comfortable with the driver state. [Add it to our CI pipeline](https://github.com/metabase/metabase-ee-extra/blob/master/.github/workflows/build.yml). Test failures: 5 Report ``` Finding tests took 29.2 s. Looking for test namespaces in directory modules/drivers/ocient/test Running 2821 tests ... Ran 2815 tests in 2169.727 seconds 14591 assertions, 5 failures, 0 errors. {:test 2815, :pass 14586, :fail 5, :error 0, :type :summary, :duration 2169726.893213, :single-threaded 2596, :parallel 219} Ran 219 tests in parallel, 2596 single-threaded. Finding and running tests took 37.3 mins. Tests failed. Failure logs: FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) 1071/2821 37% [================== ] ETA: 10:45Created a new logger for metabase.email ERROR TypeError: Cannot read property 'table.column_formatting' of null [main] TRACE metabase.email.messages - NOT including CSV attachement for Card "card-name" because less than 10 rows in results 1136/2821 40% [==================== ] ETA: 09:50Created a new logger for metabase.logger FAIL in metabase.query-processor-test.date-bucketing-test/group-by-week-test (date_bucketing_test.clj:698) JVM timezone set to Pacific :ocient expected: (["2015-05-31T00:00:00-07:00" 46] ["2015-06-07T00:00:00-07:00" 47] ["2015-06-14T00:00:00-07:00" 40] ["2015-06-21T00:00:00-07:00" 60] ["2015-06-28T00:00:00-07:00" 7]) actual: ([["2015-05-30T17:00:00-07:00" 46] ["2015-06-06T17:00:00-07:00" 47] ["2015-06-13T17:00:00-07:00" 40] ["2015-06-20T17:00:00-07:00" 60] ["2015-06-27T17:00:00-07:00" 7]]) 2185/2821 77% [====================================== ] ETA: 10:04Error in test-qp-middleware runf: #error { :cause Something went wrong :via [{:type java.lang.Exception :message Something went wrong :at [metabase.query_processor.middleware.catch_exceptions_test$fn__174855$fn__174856 invoke catch_exceptions_test.clj 85]}] ```
1.0
Review Ocient Driver - The Ocient team developed [their driver](https://github.com/Xeograph/metabase-ocient-driver) and are applying for the partnership program. To add the driver to our Cloud offerings, we need to: - [x] Code review looking for bad practices and security vulnerabilities - [x] Give them a direction to solve the failing tests - [x] As soon as we are comfortable with the driver state. [Add it to our CI pipeline](https://github.com/metabase/metabase-ee-extra/blob/master/.github/workflows/build.yml). Test failures: 5 Report ``` Finding tests took 29.2 s. Looking for test namespaces in directory modules/drivers/ocient/test Running 2821 tests ... Ran 2815 tests in 2169.727 seconds 14591 assertions, 5 failures, 0 errors. {:test 2815, :pass 14586, :fail 5, :error 0, :type :summary, :duration 2169726.893213, :single-threaded 2596, :parallel 219} Ran 219 tests in parallel, 2596 single-threaded. Finding and running tests took 37.3 mins. Tests failed. Failure logs: FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) FAIL in metabase.driver.sql-jdbc.sync.describe-database-test/database-schema-filtering-test (describe_database_test.clj:143) :ocient Filtering connections for schemas works as expected with an inclusion filter expected: (contains? #{\s \v} (first schema-name)) actual: (not (contains? #{\s \v} \m)) 1071/2821 37% [================== ] ETA: 10:45Created a new logger for metabase.email ERROR TypeError: Cannot read property 'table.column_formatting' of null [main] TRACE metabase.email.messages - NOT including CSV attachement for Card "card-name" because less than 10 rows in results 1136/2821 40% [==================== ] ETA: 09:50Created a new logger for metabase.logger FAIL in metabase.query-processor-test.date-bucketing-test/group-by-week-test (date_bucketing_test.clj:698) JVM timezone set to Pacific :ocient expected: (["2015-05-31T00:00:00-07:00" 46] ["2015-06-07T00:00:00-07:00" 47] ["2015-06-14T00:00:00-07:00" 40] ["2015-06-21T00:00:00-07:00" 60] ["2015-06-28T00:00:00-07:00" 7]) actual: ([["2015-05-30T17:00:00-07:00" 46] ["2015-06-06T17:00:00-07:00" 47] ["2015-06-13T17:00:00-07:00" 40] ["2015-06-20T17:00:00-07:00" 60] ["2015-06-27T17:00:00-07:00" 7]]) 2185/2821 77% [====================================== ] ETA: 10:04Error in test-qp-middleware runf: #error { :cause Something went wrong :via [{:type java.lang.Exception :message Something went wrong :at [metabase.query_processor.middleware.catch_exceptions_test$fn__174855$fn__174856 invoke catch_exceptions_test.clj 85]}] ```
process
review ocient driver the ocient team developed and are applying for the partnership program to add the driver to our cloud offerings we need to code review looking for bad practices and security vulnerabilities give them a direction to solve the failing tests as soon as we are comfortable with the driver state test failures report finding tests took s looking for test namespaces in directory modules drivers ocient test running tests ran tests in seconds assertions failures errors test pass fail error type summary duration single threaded parallel ran tests in parallel single threaded finding and running tests took mins tests failed failure logs fail in metabase driver sql jdbc sync describe database test database schema filtering test describe database test clj ocient filtering connections for schemas works as expected with an inclusion filter expected contains s v first schema name actual not contains s v m fail in metabase driver sql jdbc sync describe database test database schema filtering test describe database test clj ocient filtering connections for schemas works as expected with an inclusion filter expected contains s v first schema name actual not contains s v m fail in metabase driver sql jdbc sync describe database test database schema filtering test describe database test clj ocient filtering connections for schemas works as expected with an inclusion filter expected contains s v first schema name actual not contains s v m fail in metabase driver sql jdbc sync describe database test database schema filtering test describe database test clj ocient filtering connections for schemas works as expected with an inclusion filter expected contains s v first schema name actual not contains s v m eta a new logger for metabase email error typeerror cannot read property table column formatting of null trace metabase email messages not including csv attachement for card card name because less than rows in results eta a new logger for metabase logger fail in metabase query processor test date bucketing test group by week test date bucketing test clj jvm timezone set to pacific ocient expected actual eta in test qp middleware runf error cause something went wrong via type java lang exception message something went wrong at
1
176,560
28,120,526,862
IssuesEvent
2023-03-31 13:56:33
microsoft/fluentui-react-native
https://api.github.com/repos/microsoft/fluentui-react-native
closed
Windows Button incorrectly styled
Type: New Feature Area: Button Status: Blocked Priority: Low Needs: PM Input Needs: Design input Platform: Windows
According to [Fluent design website](https://developer.microsoft.com/en-us/fluentui#/controls/crossplatform/button), the Windows Button should look like this: ![image](https://user-images.githubusercontent.com/508395/82357571-2875d880-99ba-11ea-92e7-8176caa838ec.png) But currently, it looks like this: ![image](https://user-images.githubusercontent.com/508395/82357904-9ae6b880-99ba-11ea-953f-7debd0c251e5.png) Creating this issue to track this inconsistency. Please let me know if you have any questions.
1.0
Windows Button incorrectly styled - According to [Fluent design website](https://developer.microsoft.com/en-us/fluentui#/controls/crossplatform/button), the Windows Button should look like this: ![image](https://user-images.githubusercontent.com/508395/82357571-2875d880-99ba-11ea-92e7-8176caa838ec.png) But currently, it looks like this: ![image](https://user-images.githubusercontent.com/508395/82357904-9ae6b880-99ba-11ea-953f-7debd0c251e5.png) Creating this issue to track this inconsistency. Please let me know if you have any questions.
non_process
windows button incorrectly styled according to the windows button should look like this but currently it looks like this creating this issue to track this inconsistency please let me know if you have any questions
0
62,871
26,192,571,008
IssuesEvent
2023-01-03 10:25:47
wso2/apk
https://api.github.com/repos/wso2/apk
closed
[Admin DS] Need to properly handle defaultLimit response data in Application Rate Plans and in Business Plans REST APIs
component/admin-domain-services
**Description:** Response for POST call on {{adminAPIServerBaseURL}}/api/am/admin/application-rate-plans { "defaultLimit": { "type": "ApplicationThrottlePolicy", "requestCount": { "requestCount": 95, "timeUnit": "min", "unitTime": 1 } }, "planId": "01ed8148-1ba8-15b6-b40e-e19e79c5d6a9", "planName": "95PerMinPlan1", "displayName": "95 Requests per Minutes", "description": "95 Requests per Minutes", "isDeployed": false } Response from {{adminAPIServerBaseURL}}/api/am/admin/application-rate-plans/{{applicationRatePlanID}} { "defaultLimit": null, "planId": "01ed8148-1ba8-15b6-b40e-e19e79c5d6a9", "planName": "95PerMinPlan1", "displayName": "95 Requests per Minutes", "description": "95 Requests per Minutes", "isDeployed": false, "type": "ApplicationThrottlePolicy", "requestcount": 95, "timeunit": "min", "unittime": 1 } As you can see above responses are different ideally they should return same. Also in the 2nd response we have not followed camel case "requestcount": 95, "timeunit": "min", "unittime": 1 Lets fix these.
1.0
[Admin DS] Need to properly handle defaultLimit response data in Application Rate Plans and in Business Plans REST APIs - **Description:** Response for POST call on {{adminAPIServerBaseURL}}/api/am/admin/application-rate-plans { "defaultLimit": { "type": "ApplicationThrottlePolicy", "requestCount": { "requestCount": 95, "timeUnit": "min", "unitTime": 1 } }, "planId": "01ed8148-1ba8-15b6-b40e-e19e79c5d6a9", "planName": "95PerMinPlan1", "displayName": "95 Requests per Minutes", "description": "95 Requests per Minutes", "isDeployed": false } Response from {{adminAPIServerBaseURL}}/api/am/admin/application-rate-plans/{{applicationRatePlanID}} { "defaultLimit": null, "planId": "01ed8148-1ba8-15b6-b40e-e19e79c5d6a9", "planName": "95PerMinPlan1", "displayName": "95 Requests per Minutes", "description": "95 Requests per Minutes", "isDeployed": false, "type": "ApplicationThrottlePolicy", "requestcount": 95, "timeunit": "min", "unittime": 1 } As you can see above responses are different ideally they should return same. Also in the 2nd response we have not followed camel case "requestcount": 95, "timeunit": "min", "unittime": 1 Lets fix these.
non_process
need to properly handle defaultlimit response data in application rate plans and in business plans rest apis description response for post call on adminapiserverbaseurl api am admin application rate plans defaultlimit type applicationthrottlepolicy requestcount requestcount timeunit min unittime planid planname displayname requests per minutes description requests per minutes isdeployed false response from adminapiserverbaseurl api am admin application rate plans applicationrateplanid defaultlimit null planid planname displayname requests per minutes description requests per minutes isdeployed false type applicationthrottlepolicy requestcount timeunit min unittime as you can see above responses are different ideally they should return same also in the response we have not followed camel case requestcount timeunit min unittime lets fix these
0
452,568
13,055,894,852
IssuesEvent
2020-07-30 03:02:57
kubesphere/console
https://api.github.com/repos/kubesphere/console
closed
Cannot create credentials from creating pipeline dialog
area/devops kind/bug kind/need-to-verify priority/high
Create button is not working. ![image](https://user-images.githubusercontent.com/22350668/88474970-49c8e700-cf5e-11ea-8d71-97cfa746a3f2.png) ![image](https://user-images.githubusercontent.com/22350668/88474988-80066680-cf5e-11ea-9aed-c552424c0bcc.png)
1.0
Cannot create credentials from creating pipeline dialog - Create button is not working. ![image](https://user-images.githubusercontent.com/22350668/88474970-49c8e700-cf5e-11ea-8d71-97cfa746a3f2.png) ![image](https://user-images.githubusercontent.com/22350668/88474988-80066680-cf5e-11ea-9aed-c552424c0bcc.png)
non_process
cannot create credentials from creating pipeline dialog create button is not working
0
675,083
23,078,371,367
IssuesEvent
2022-07-26 03:32:11
p5-serial/p5.serialcontrol
https://api.github.com/repos/p5-serial/p5.serialcontrol
closed
Add aria and labels to interface
enhancement high priority
aria live for list ports, labels to checkbox and other form elements
1.0
Add aria and labels to interface - aria live for list ports, labels to checkbox and other form elements
non_process
add aria and labels to interface aria live for list ports labels to checkbox and other form elements
0
1,665
4,295,277,623
IssuesEvent
2016-07-19 06:10:37
ParsePlatform/parse-server
https://api.github.com/repos/ParsePlatform/parse-server
closed
_email_verify_token never expires blocking new signups...
in-process pr-submitted
Describe your issue in as much detail as possible. The following are some cases where the parse-server blocks `new` user Sign Ups because the `_email_verify_token` never expires. ### Steps to reproduce Scenario 1: User enters an email address with a typo during the signup process preventing another user from signing up. 1. App requires `username` to be the same as the `email address` 2. `User1` has an email address `johnsmith111@gmail.com` 3. `User2` has an email address `johnsmith1@gmail.com` 4. `User1` enters `johnsmith1@gmail.com` in the email address field during sign up - clearly a typo. User1 should have entered: `johnsmith111@gmail.com` - `User2` gets the email address verification - `User2` deletes the email as they did not initiate the signup 5. `User2` now tries to signup for the same service at a later time (lets say after 24 hours) - `User2` enters `johnsmith111@gmail.com` in the email field #### Expected Results - App should allow `User2` to sign up for the service - Parse Server should `not` allow the user to reuse the email verification link after successfully signing up for the service - Parse Server should delete the `_email_verify_token` if sign up succeeds. #### Actual Outcome - **App returns an ERROR saying that this `username` is taken** - Parse Server allows the user to click on the email verification link as many times as they want `even` after the user has successfully signed up for the service - Parse Server does not delete the `_email_verify_token` after Sign Up is successful. ### Potential solution 1. Parse Server should provide additional config variables like `allowEmailVerifyTokenToExpire` and `emailVerifyTokenValidityDuration` 2. Parse Server should default `emailVerifyTokenValidityDuration` to `1 year` for backward compatibility with existing apps - Developers should typically set this to 1 hour or 2 hours as a best practice ##### if `allowEmailVerifyTokenToExpire` and `verifyUserEmails` are set to `true` then 3. Parse Server should save `_email_verify_token_expires_at` in the _User class along with the `_email_verify_token` at `SignUp` time - `_email_verify_token_expires_at` = `currentTime` + `emailVerifyTokenValidityDuration` 4. If the user clicks on the email verification link `BEFORE` the `_email_verify_token_expires_at` then - update `emailVerified` to `true` in the `User` class (this code is already in place) - delete the `_email_verify_token` and `_email_verify_token_expires_at` fields for this user (new code) - respond back with a success (this code is already in place) 5. If the user clicks on the email verification link `AFTER` the `_email_verify_token_expires_at` then - respond back with an error: `email verification token expired.` - user can try again and create a new account 6. If the user completed `email verification` successfully but clicks on the link for second time then return an error message (i.e., fail gracefully) 7. Developer can also create a scheduled `clean up` job. This job can run periodically (say every 5 minutes) to do the following: - find users with - `emailVerified` set to `false` **(email not verified)** - and `_email_verify_token_expires_at` less than `current time` **(token expired)** - delete the users found in the above step ### Environment Setup - **Server** - parse-server version: 2.2.15 - Operating System: OSX - Hardware: Macbook - Localhost or remote server? (AWS, Heroku, Azure, Digital Ocean, etc): Localhost - **Database** - MongoDB version: 3.2.5 - Storage engine: N/A - Hardware: Macbook - Localhost or remote server? (AWS, mLab, ObjectRocket, Digital Ocean, etc): Localhost - As you can see the email addresses are pretty similar to each other - There are many other cases where email addresses out there are similar to each other. for example: `johnsmith@gmail.com`, `john.smith@gmail.com`, `johnsmith@yahoo.com` and so on... - When email addresses are similar, there is chance for the users to type in the wrong email address during sign up (a honest typo)
1.0
_email_verify_token never expires blocking new signups... - Describe your issue in as much detail as possible. The following are some cases where the parse-server blocks `new` user Sign Ups because the `_email_verify_token` never expires. ### Steps to reproduce Scenario 1: User enters an email address with a typo during the signup process preventing another user from signing up. 1. App requires `username` to be the same as the `email address` 2. `User1` has an email address `johnsmith111@gmail.com` 3. `User2` has an email address `johnsmith1@gmail.com` 4. `User1` enters `johnsmith1@gmail.com` in the email address field during sign up - clearly a typo. User1 should have entered: `johnsmith111@gmail.com` - `User2` gets the email address verification - `User2` deletes the email as they did not initiate the signup 5. `User2` now tries to signup for the same service at a later time (lets say after 24 hours) - `User2` enters `johnsmith111@gmail.com` in the email field #### Expected Results - App should allow `User2` to sign up for the service - Parse Server should `not` allow the user to reuse the email verification link after successfully signing up for the service - Parse Server should delete the `_email_verify_token` if sign up succeeds. #### Actual Outcome - **App returns an ERROR saying that this `username` is taken** - Parse Server allows the user to click on the email verification link as many times as they want `even` after the user has successfully signed up for the service - Parse Server does not delete the `_email_verify_token` after Sign Up is successful. ### Potential solution 1. Parse Server should provide additional config variables like `allowEmailVerifyTokenToExpire` and `emailVerifyTokenValidityDuration` 2. Parse Server should default `emailVerifyTokenValidityDuration` to `1 year` for backward compatibility with existing apps - Developers should typically set this to 1 hour or 2 hours as a best practice ##### if `allowEmailVerifyTokenToExpire` and `verifyUserEmails` are set to `true` then 3. Parse Server should save `_email_verify_token_expires_at` in the _User class along with the `_email_verify_token` at `SignUp` time - `_email_verify_token_expires_at` = `currentTime` + `emailVerifyTokenValidityDuration` 4. If the user clicks on the email verification link `BEFORE` the `_email_verify_token_expires_at` then - update `emailVerified` to `true` in the `User` class (this code is already in place) - delete the `_email_verify_token` and `_email_verify_token_expires_at` fields for this user (new code) - respond back with a success (this code is already in place) 5. If the user clicks on the email verification link `AFTER` the `_email_verify_token_expires_at` then - respond back with an error: `email verification token expired.` - user can try again and create a new account 6. If the user completed `email verification` successfully but clicks on the link for second time then return an error message (i.e., fail gracefully) 7. Developer can also create a scheduled `clean up` job. This job can run periodically (say every 5 minutes) to do the following: - find users with - `emailVerified` set to `false` **(email not verified)** - and `_email_verify_token_expires_at` less than `current time` **(token expired)** - delete the users found in the above step ### Environment Setup - **Server** - parse-server version: 2.2.15 - Operating System: OSX - Hardware: Macbook - Localhost or remote server? (AWS, Heroku, Azure, Digital Ocean, etc): Localhost - **Database** - MongoDB version: 3.2.5 - Storage engine: N/A - Hardware: Macbook - Localhost or remote server? (AWS, mLab, ObjectRocket, Digital Ocean, etc): Localhost - As you can see the email addresses are pretty similar to each other - There are many other cases where email addresses out there are similar to each other. for example: `johnsmith@gmail.com`, `john.smith@gmail.com`, `johnsmith@yahoo.com` and so on... - When email addresses are similar, there is chance for the users to type in the wrong email address during sign up (a honest typo)
process
email verify token never expires blocking new signups describe your issue in as much detail as possible the following are some cases where the parse server blocks new user sign ups because the email verify token never expires steps to reproduce scenario user enters an email address with a typo during the signup process preventing another user from signing up app requires username to be the same as the email address has an email address gmail com has an email address gmail com enters gmail com in the email address field during sign up clearly a typo should have entered gmail com gets the email address verification deletes the email as they did not initiate the signup now tries to signup for the same service at a later time lets say after hours enters gmail com in the email field expected results app should allow to sign up for the service parse server should not allow the user to reuse the email verification link after successfully signing up for the service parse server should delete the email verify token if sign up succeeds actual outcome app returns an error saying that this username is taken parse server allows the user to click on the email verification link as many times as they want even after the user has successfully signed up for the service parse server does not delete the email verify token after sign up is successful potential solution parse server should provide additional config variables like allowemailverifytokentoexpire and emailverifytokenvalidityduration parse server should default emailverifytokenvalidityduration to year for backward compatibility with existing apps developers should typically set this to hour or hours as a best practice if allowemailverifytokentoexpire and verifyuseremails are set to true then parse server should save email verify token expires at in the user class along with the email verify token at signup time email verify token expires at currenttime emailverifytokenvalidityduration if the user clicks on the email verification link before the email verify token expires at then update emailverified to true in the user class this code is already in place delete the email verify token and email verify token expires at fields for this user new code respond back with a success this code is already in place if the user clicks on the email verification link after the email verify token expires at then respond back with an error email verification token expired user can try again and create a new account if the user completed email verification successfully but clicks on the link for second time then return an error message i e fail gracefully developer can also create a scheduled clean up job this job can run periodically say every minutes to do the following find users with emailverified set to false email not verified and email verify token expires at less than current time token expired delete the users found in the above step environment setup server parse server version operating system osx hardware macbook localhost or remote server aws heroku azure digital ocean etc localhost database mongodb version storage engine n a hardware macbook localhost or remote server aws mlab objectrocket digital ocean etc localhost as you can see the email addresses are pretty similar to each other there are many other cases where email addresses out there are similar to each other for example johnsmith gmail com john smith gmail com johnsmith yahoo com and so on when email addresses are similar there is chance for the users to type in the wrong email address during sign up a honest typo
1
21,694
30,190,785,401
IssuesEvent
2023-07-04 15:12:52
zammad/zammad
https://api.github.com/repos/zammad/zammad
opened
Missing spaces in signatures created with Exclaimer
bug verified mail processing signature
### Used Zammad Version 6.0 ### Environment - Installation method: any - Operating system: any - Database + version: any - Elasticsearch version: any - Browser + version: any ### Actual behaviour When receiving emails with signatures that have been crated with Exclaimer (exclaimer.com), in some cases, spaces are omitted. ![image](https://github.com/zammad/zammad/assets/69846837/988d3aef-2762-4281-9850-ff53ca957f60) ### Expected behaviour The signature should be represented correctly, including all spaces. ### Steps to reproduce the behaviour * Create a signature using Exclaimer. * Email a Zammad instance. Please see the attached example text file (renamed from .eml). [ticket-example.txt](https://github.com/zammad/zammad/files/11950600/ticket-example.txt) ### Support Ticket Ticket#10132104 ### I'm sure this is a bug and no feature request or a general question. yes
1.0
Missing spaces in signatures created with Exclaimer - ### Used Zammad Version 6.0 ### Environment - Installation method: any - Operating system: any - Database + version: any - Elasticsearch version: any - Browser + version: any ### Actual behaviour When receiving emails with signatures that have been crated with Exclaimer (exclaimer.com), in some cases, spaces are omitted. ![image](https://github.com/zammad/zammad/assets/69846837/988d3aef-2762-4281-9850-ff53ca957f60) ### Expected behaviour The signature should be represented correctly, including all spaces. ### Steps to reproduce the behaviour * Create a signature using Exclaimer. * Email a Zammad instance. Please see the attached example text file (renamed from .eml). [ticket-example.txt](https://github.com/zammad/zammad/files/11950600/ticket-example.txt) ### Support Ticket Ticket#10132104 ### I'm sure this is a bug and no feature request or a general question. yes
process
missing spaces in signatures created with exclaimer used zammad version environment installation method any operating system any database version any elasticsearch version any browser version any actual behaviour when receiving emails with signatures that have been crated with exclaimer exclaimer com in some cases spaces are omitted expected behaviour the signature should be represented correctly including all spaces steps to reproduce the behaviour create a signature using exclaimer email a zammad instance please see the attached example text file renamed from eml support ticket ticket i m sure this is a bug and no feature request or a general question yes
1
103,304
22,284,662,927
IssuesEvent
2022-06-11 12:29:18
arnauddupuis/pygamelib
https://api.github.com/repos/arnauddupuis/pygamelib
closed
Improve RandomActuator
enhancement good first issue new coders friendly
**Problem summary/missing feature:** The RandomActuator class is currently picking a new direction each turn. This translate into a very poor behavior were it looks like the NPC is having a seizure. **Expected behavior:** Ideally, the RandomActuator should pick a direction and a number of steps randomly and go into that direction for the number of steps. When steps reaches 0, pick a new direction and number of steps and do it again. Additionally, the direction is returned as a constant (UP, DOWN, etc.) but it would probably be better to return a Vector2D. **Work to do:** In `RandomActuator`, add 2 member variables: direction and steps. In `RandomActuator.next_move()` check if steps is 0, in that case choose a new direction and a new number of steps randomly. Return the new direction and decrease steps. If steps is not 0 return the direction and decrease steps. `direction` should be a Vector2D with each component being either -1, 0 or 1. Bonus: If the NPC position hasn't change in 2 turns, it is blocked so choose a new direction.
1.0
Improve RandomActuator - **Problem summary/missing feature:** The RandomActuator class is currently picking a new direction each turn. This translate into a very poor behavior were it looks like the NPC is having a seizure. **Expected behavior:** Ideally, the RandomActuator should pick a direction and a number of steps randomly and go into that direction for the number of steps. When steps reaches 0, pick a new direction and number of steps and do it again. Additionally, the direction is returned as a constant (UP, DOWN, etc.) but it would probably be better to return a Vector2D. **Work to do:** In `RandomActuator`, add 2 member variables: direction and steps. In `RandomActuator.next_move()` check if steps is 0, in that case choose a new direction and a new number of steps randomly. Return the new direction and decrease steps. If steps is not 0 return the direction and decrease steps. `direction` should be a Vector2D with each component being either -1, 0 or 1. Bonus: If the NPC position hasn't change in 2 turns, it is blocked so choose a new direction.
non_process
improve randomactuator problem summary missing feature the randomactuator class is currently picking a new direction each turn this translate into a very poor behavior were it looks like the npc is having a seizure expected behavior ideally the randomactuator should pick a direction and a number of steps randomly and go into that direction for the number of steps when steps reaches pick a new direction and number of steps and do it again additionally the direction is returned as a constant up down etc but it would probably be better to return a work to do in randomactuator add member variables direction and steps in randomactuator next move check if steps is in that case choose a new direction and a new number of steps randomly return the new direction and decrease steps if steps is not return the direction and decrease steps direction should be a with each component being either or bonus if the npc position hasn t change in turns it is blocked so choose a new direction
0
14,028
16,827,221,047
IssuesEvent
2021-06-17 20:21:34
googleapis/python-vision
https://api.github.com/repos/googleapis/python-vision
closed
TextAnnotation doesn't have orientationDegrees
api: vision type: process
REST v1 AnnotateImageResponse already returns `orientationDegrees` https://cloud.google.com/vision/docs/reference/rest/v1/AnnotateImageResponse#Page but looks like this library proto files hasn't been updated for a while: https://github.com/googleapis/googleapis/blob/master/google/cloud/vision/v1/text_annotation.proto#L96 https://github.com/googleapis/googleapis/blob/master/google/cloud/vision/v1p4beta1/text_annotation.proto#L98 #### Environment details - `google-cloud-vision` version: 2.3.0 #### Steps to reproduce 1. Send any DOCUMENT_TEXT_DETECTION request to REST \ with this library 2. check `Pages` A similar issue https://github.com/googleapis/python-vision/issues/138
1.0
TextAnnotation doesn't have orientationDegrees - REST v1 AnnotateImageResponse already returns `orientationDegrees` https://cloud.google.com/vision/docs/reference/rest/v1/AnnotateImageResponse#Page but looks like this library proto files hasn't been updated for a while: https://github.com/googleapis/googleapis/blob/master/google/cloud/vision/v1/text_annotation.proto#L96 https://github.com/googleapis/googleapis/blob/master/google/cloud/vision/v1p4beta1/text_annotation.proto#L98 #### Environment details - `google-cloud-vision` version: 2.3.0 #### Steps to reproduce 1. Send any DOCUMENT_TEXT_DETECTION request to REST \ with this library 2. check `Pages` A similar issue https://github.com/googleapis/python-vision/issues/138
process
textannotation doesn t have orientationdegrees rest annotateimageresponse already returns orientationdegrees but looks like this library proto files hasn t been updated for a while environment details google cloud vision version steps to reproduce send any document text detection request to rest with this library check pages a similar issue
1
5,842
8,667,017,763
IssuesEvent
2018-11-29 07:06:53
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Add process.memoryUsage().toString for human-readable output
process
**Problem**: result of `process.memoryUsage()` is hardly readable by humans, limiting it's usability for fast debugging. **Proposed solution**: Add `toString` method to `process.memoryUsage` results. **Behaviour now**: `process.memoryUsage().toString() === "[object Object]"` **Expected behavior**: ```javascript /* { rss: 29233152, heapTotal: 7159808, heapUsed: 4383040, external: 8224 } */ process.memoryUsage().toString(); // "rss: 27.9M; heapTotal: 6.8M; heapUsed: 4.2M; external: 8K" ```
1.0
Add process.memoryUsage().toString for human-readable output - **Problem**: result of `process.memoryUsage()` is hardly readable by humans, limiting it's usability for fast debugging. **Proposed solution**: Add `toString` method to `process.memoryUsage` results. **Behaviour now**: `process.memoryUsage().toString() === "[object Object]"` **Expected behavior**: ```javascript /* { rss: 29233152, heapTotal: 7159808, heapUsed: 4383040, external: 8224 } */ process.memoryUsage().toString(); // "rss: 27.9M; heapTotal: 6.8M; heapUsed: 4.2M; external: 8K" ```
process
add process memoryusage tostring for human readable output problem result of process memoryusage is hardly readable by humans limiting it s usability for fast debugging proposed solution add tostring method to process memoryusage results behaviour now process memoryusage tostring expected behavior javascript rss heaptotal heapused external process memoryusage tostring rss heaptotal heapused external
1
583,710
17,396,775,806
IssuesEvent
2021-08-02 14:21:00
wwosimulation/tracker
https://api.github.com/repos/wwosimulation/tracker
reopened
DB failing to initialize players
Bug Config Economy High Priority
`+xp add MaximRq 250` command doesn't works/responds ** ** The above bug was reported by inno on ice#0419 User ID: 719564153072910407 Location: #commands (606123748738859008) in Wolvesville Simulation
1.0
DB failing to initialize players - `+xp add MaximRq 250` command doesn't works/responds ** ** The above bug was reported by inno on ice#0419 User ID: 719564153072910407 Location: #commands (606123748738859008) in Wolvesville Simulation
non_process
db failing to initialize players xp add maximrq command doesn t works responds the above bug was reported by inno on ice user id location commands in wolvesville simulation
0
18,504
24,551,276,487
IssuesEvent
2022-10-12 12:48:03
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] [Edge browser] Sign in screen > Password field > 'Eye' icon is not getting displayed in the following scenario
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
Steps: 1. Open the edge browser 2. Enter the PM test URL 3. Enter a valid email id 4. Click on the Password field and enter the password 5. Click on somewhere on the screen 6. Again click on the Password field and observe AR: 'Eye' icon is not getting displayed ER: 'Eye' icon should get displayed to the user
3.0
[PM] [Edge browser] Sign in screen > Password field > 'Eye' icon is not getting displayed in the following scenario - Steps: 1. Open the edge browser 2. Enter the PM test URL 3. Enter a valid email id 4. Click on the Password field and enter the password 5. Click on somewhere on the screen 6. Again click on the Password field and observe AR: 'Eye' icon is not getting displayed ER: 'Eye' icon should get displayed to the user
process
sign in screen password field eye icon is not getting displayed in the following scenario steps open the edge browser enter the pm test url enter a valid email id click on the password field and enter the password click on somewhere on the screen again click on the password field and observe ar eye icon is not getting displayed er eye icon should get displayed to the user
1
102,456
12,799,157,266
IssuesEvent
2020-07-02 14:58:44
uncharted-distil/distil
https://api.github.com/repos/uncharted-distil/distil
closed
Compound facet creation
design
Creation of timeseries and geo-coordinate facets are currently still in the prototype form. A proper UI needs to be designed for both of these.
1.0
Compound facet creation - Creation of timeseries and geo-coordinate facets are currently still in the prototype form. A proper UI needs to be designed for both of these.
non_process
compound facet creation creation of timeseries and geo coordinate facets are currently still in the prototype form a proper ui needs to be designed for both of these
0
8,485
11,645,068,193
IssuesEvent
2020-02-29 22:33:56
googleapis/nodejs-text-to-speech
https://api.github.com/repos/googleapis/nodejs-text-to-speech
closed
GA release
api: texttospeech type: process
Package name: `@google-cloud/text-to-speech` Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [x] Most common / important scenarios have descriptive samples - [x] Public manual methods have at least one usage sample each (excluding overloads) - [x] Per-API README includes a full description of the API - [x] Per-API README contains at least one “getting started” sample using the most common API scenario - [x] Manual code has been reviewed by API producer - [x] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
GA release - Package name: `@google-cloud/text-to-speech` Current release: **beta** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [x] Most common / important scenarios have descriptive samples - [x] Public manual methods have at least one usage sample each (excluding overloads) - [x] Per-API README includes a full description of the API - [x] Per-API README contains at least one “getting started” sample using the most common API scenario - [x] Manual code has been reviewed by API producer - [x] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
ga release package name google cloud text to speech current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
13,169
15,594,507,211
IssuesEvent
2021-03-18 13:58:52
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
MAINT Move PolynomialFeatures to _polynomial.py
Easy module:preprocessing
A little maintenance task: Move `class PolynomialFeatures` from `_data.py` to `_polynomial.py` and tests likewise. See https://github.com/scikit-learn/scikit-learn/pull/18368#issuecomment-758319583.
1.0
MAINT Move PolynomialFeatures to _polynomial.py - A little maintenance task: Move `class PolynomialFeatures` from `_data.py` to `_polynomial.py` and tests likewise. See https://github.com/scikit-learn/scikit-learn/pull/18368#issuecomment-758319583.
process
maint move polynomialfeatures to polynomial py a little maintenance task move class polynomialfeatures from data py to polynomial py and tests likewise see
1
7,793
10,948,817,351
IssuesEvent
2019-11-26 09:38:58
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Altcoins Instant trade period is wrong
an:investigation in:trade-process was:dropped
Trade period should start after deposit transaction get 1 confimation, correct? Instead trade period starts when offer was taken, so if deposit confirms 40 minutes after offer was taken, traders only have 20 minutes to complete trade. Sometimes very hard /impossible when fee jumps. Notice in after screenshot: end of trade period (8:44pm) is exactly 1 hour after offer was taken (7:44pm). ![before-confirmation-pixelated](https://user-images.githubusercontent.com/51380215/63352847-59ed4500-c351-11e9-957b-607aca26a57f.png) ![after-confirmation-pixelated](https://user-images.githubusercontent.com/51380215/63352844-5954ae80-c351-11e9-9f01-410ed668f86c.png)
1.0
Altcoins Instant trade period is wrong - Trade period should start after deposit transaction get 1 confimation, correct? Instead trade period starts when offer was taken, so if deposit confirms 40 minutes after offer was taken, traders only have 20 minutes to complete trade. Sometimes very hard /impossible when fee jumps. Notice in after screenshot: end of trade period (8:44pm) is exactly 1 hour after offer was taken (7:44pm). ![before-confirmation-pixelated](https://user-images.githubusercontent.com/51380215/63352847-59ed4500-c351-11e9-957b-607aca26a57f.png) ![after-confirmation-pixelated](https://user-images.githubusercontent.com/51380215/63352844-5954ae80-c351-11e9-9f01-410ed668f86c.png)
process
altcoins instant trade period is wrong trade period should start after deposit transaction get confimation correct instead trade period starts when offer was taken so if deposit confirms minutes after offer was taken traders only have minutes to complete trade sometimes very hard impossible when fee jumps notice in after screenshot end of trade period is exactly hour after offer was taken
1
286,406
31,563,476,476
IssuesEvent
2023-09-03 14:30:13
ZSBRybnik/backend
https://api.github.com/repos/ZSBRybnik/backend
closed
scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl: 2 vulnerabilities (highest severity is: 9.8) - autoclosed
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p></summary> <p>SciPy: Scientific Library for Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /source/native-addon-python/requirements.txt</p> <p>Path to vulnerable library: /source/native-addon-python/requirements.txt,/source/native-addon-python/requirements.txt</p> <p> <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/backend/commit/d6f11282ee9f4af0baad5df06570b6a9700994af">d6f11282ee9f4af0baad5df06570b6a9700994af</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (scipy version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2023-29824](https://www.mend.io/vulnerability-database/CVE-2023-29824) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl | Direct | 1.8.0 | &#10060; | | [CVE-2023-25399](https://www.mend.io/vulnerability-database/CVE-2023-25399) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.5 | scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl | Direct | 1.10.0 | &#10060; | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2023-29824</summary> ### Vulnerable Library - <b>scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p> <p>SciPy: Scientific Library for Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /source/native-addon-python/requirements.txt</p> <p>Path to vulnerable library: /source/native-addon-python/requirements.txt,/source/native-addon-python/requirements.txt</p> <p> Dependency Hierarchy: - :x: **scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/backend/commit/d6f11282ee9f4af0baad5df06570b6a9700994af">d6f11282ee9f4af0baad5df06570b6a9700994af</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> ** DISPUTED ** A use-after-free issue was discovered in Py_FindObjects() function in SciPy versions prior to 1.8.0. NOTE: the vendor and discoverer indicate that this is not a security issue. <p>Publish Date: 2023-07-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-29824>CVE-2023-29824</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-29824">https://www.cve.org/CVERecord?id=CVE-2023-29824</a></p> <p>Release Date: 2023-07-06</p> <p>Fix Resolution: 1.8.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2023-25399</summary> ### Vulnerable Library - <b>scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p> <p>SciPy: Scientific Library for Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /source/native-addon-python/requirements.txt</p> <p>Path to vulnerable library: /source/native-addon-python/requirements.txt,/source/native-addon-python/requirements.txt</p> <p> Dependency Hierarchy: - :x: **scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/backend/commit/d6f11282ee9f4af0baad5df06570b6a9700994af">d6f11282ee9f4af0baad5df06570b6a9700994af</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> A refcounting issue which leads to potential memory leak was discovered in scipy commit 8627df31ab in Py_FindObjects() function. <p>Publish Date: 2023-07-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25399>CVE-2023-25399</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25399">https://www.cve.org/CVERecord?id=CVE-2023-25399</a></p> <p>Release Date: 2023-07-05</p> <p>Fix Resolution: 1.10.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl: 2 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p></summary> <p>SciPy: Scientific Library for Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /source/native-addon-python/requirements.txt</p> <p>Path to vulnerable library: /source/native-addon-python/requirements.txt,/source/native-addon-python/requirements.txt</p> <p> <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/backend/commit/d6f11282ee9f4af0baad5df06570b6a9700994af">d6f11282ee9f4af0baad5df06570b6a9700994af</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (scipy version) | Remediation Possible** | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2023-29824](https://www.mend.io/vulnerability-database/CVE-2023-29824) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl | Direct | 1.8.0 | &#10060; | | [CVE-2023-25399](https://www.mend.io/vulnerability-database/CVE-2023-25399) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.5 | scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl | Direct | 1.10.0 | &#10060; | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2023-29824</summary> ### Vulnerable Library - <b>scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p> <p>SciPy: Scientific Library for Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /source/native-addon-python/requirements.txt</p> <p>Path to vulnerable library: /source/native-addon-python/requirements.txt,/source/native-addon-python/requirements.txt</p> <p> Dependency Hierarchy: - :x: **scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/backend/commit/d6f11282ee9f4af0baad5df06570b6a9700994af">d6f11282ee9f4af0baad5df06570b6a9700994af</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> ** DISPUTED ** A use-after-free issue was discovered in Py_FindObjects() function in SciPy versions prior to 1.8.0. NOTE: the vendor and discoverer indicate that this is not a security issue. <p>Publish Date: 2023-07-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-29824>CVE-2023-29824</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-29824">https://www.cve.org/CVERecord?id=CVE-2023-29824</a></p> <p>Release Date: 2023-07-06</p> <p>Fix Resolution: 1.8.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2023-25399</summary> ### Vulnerable Library - <b>scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</b></p> <p>SciPy: Scientific Library for Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/58/4f/11f34cfc57ead25752a7992b069c36f5d18421958ebd6466ecd849aeaf86/scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /source/native-addon-python/requirements.txt</p> <p>Path to vulnerable library: /source/native-addon-python/requirements.txt,/source/native-addon-python/requirements.txt</p> <p> Dependency Hierarchy: - :x: **scipy-1.7.3-cp37-cp37m-manylinux_2_12_x86_64.manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/backend/commit/d6f11282ee9f4af0baad5df06570b6a9700994af">d6f11282ee9f4af0baad5df06570b6a9700994af</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> A refcounting issue which leads to potential memory leak was discovered in scipy commit 8627df31ab in Py_FindObjects() function. <p>Publish Date: 2023-07-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25399>CVE-2023-25399</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25399">https://www.cve.org/CVERecord?id=CVE-2023-25399</a></p> <p>Release Date: 2023-07-05</p> <p>Fix Resolution: 1.10.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_process
scipy manylinux whl vulnerabilities highest severity is autoclosed vulnerable library scipy manylinux whl scipy scientific library for python library home page a href path to dependency file source native addon python requirements txt path to vulnerable library source native addon python requirements txt source native addon python requirements txt found in head commit a href vulnerabilities cve severity cvss dependency type fixed in scipy version remediation possible critical scipy manylinux whl direct medium scipy manylinux whl direct in some cases remediation pr cannot be created automatically for a vulnerability despite the availability of remediation details cve vulnerable library scipy manylinux whl scipy scientific library for python library home page a href path to dependency file source native addon python requirements txt path to vulnerable library source native addon python requirements txt source native addon python requirements txt dependency hierarchy x scipy manylinux whl vulnerable library found in head commit a href found in base branch master vulnerability details disputed a use after free issue was discovered in py findobjects function in scipy versions prior to note the vendor and discoverer indicate that this is not a security issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend cve vulnerable library scipy manylinux whl scipy scientific library for python library home page a href path to dependency file source native addon python requirements txt path to vulnerable library source native addon python requirements txt source native addon python requirements txt dependency hierarchy x scipy manylinux whl vulnerable library found in head commit a href found in base branch master vulnerability details a refcounting issue which leads to potential memory leak was discovered in scipy commit in py findobjects function publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
12,521
14,966,737,685
IssuesEvent
2021-01-27 14:55:30
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
closed
Schema definitions for 'Native' log types
p1 team:data processing
We need to have a way to express our 'Native' parsers and schemas in YAML so they can be part of the Managed Schemas repository. - Add a `native` parser config in `logschema.Schema` that will delegate the parsing for a Schema to the relevant `Go` parser. - Add a `logschema.FromStruct` function to convert the Event Schemas defined as `Go` structs to `logschema.Schema` - Have a `logtypes.NativeConfig` entry builder to expose the 'Managed' schemas as log types.
1.0
Schema definitions for 'Native' log types - We need to have a way to express our 'Native' parsers and schemas in YAML so they can be part of the Managed Schemas repository. - Add a `native` parser config in `logschema.Schema` that will delegate the parsing for a Schema to the relevant `Go` parser. - Add a `logschema.FromStruct` function to convert the Event Schemas defined as `Go` structs to `logschema.Schema` - Have a `logtypes.NativeConfig` entry builder to expose the 'Managed' schemas as log types.
process
schema definitions for native log types we need to have a way to express our native parsers and schemas in yaml so they can be part of the managed schemas repository add a native parser config in logschema schema that will delegate the parsing for a schema to the relevant go parser add a logschema fromstruct function to convert the event schemas defined as go structs to logschema schema have a logtypes nativeconfig entry builder to expose the managed schemas as log types
1
1,858
4,680,471,824
IssuesEvent
2016-10-08 07:18:47
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Investigate flaky parallel/test-child-process-fork-regr-gh-2847
child_process test
* **Version**: master * **Platform**: aix * **Subsystem**: child_process This test failure recently occurred on AIX: https://ci.nodejs.org/job/node-test-commit-aix/1317/nodes=aix61-ppc64/console Output: ``` not ok 74 parallel/test-child-process-fork-regr-gh-2847 # /home/iojs/build/workspace/node-test-commit-aix/nodes/aix61-ppc64/test/parallel/test-child-process-fork-regr-gh-2847.js:58 # throw err; # ^ # # Error: channel closed # at ChildProcess.target.send (internal/child_process.js:540:16) # at Worker.send (cluster.js:65:28) # at Socket.<anonymous> (/home/iojs/build/workspace/node-test-commit-aix/nodes/aix61-ppc64/test/parallel/test-child-process-fork-regr-gh-2847.js:33:14) # at Socket.g (events.js:291:16) # at emitNone (events.js:86:13) # at Socket.emit (events.js:185:7) # at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1075:10) --- duration_ms: 0.380 ``` /cc @mhdawson ?
1.0
Investigate flaky parallel/test-child-process-fork-regr-gh-2847 - * **Version**: master * **Platform**: aix * **Subsystem**: child_process This test failure recently occurred on AIX: https://ci.nodejs.org/job/node-test-commit-aix/1317/nodes=aix61-ppc64/console Output: ``` not ok 74 parallel/test-child-process-fork-regr-gh-2847 # /home/iojs/build/workspace/node-test-commit-aix/nodes/aix61-ppc64/test/parallel/test-child-process-fork-regr-gh-2847.js:58 # throw err; # ^ # # Error: channel closed # at ChildProcess.target.send (internal/child_process.js:540:16) # at Worker.send (cluster.js:65:28) # at Socket.<anonymous> (/home/iojs/build/workspace/node-test-commit-aix/nodes/aix61-ppc64/test/parallel/test-child-process-fork-regr-gh-2847.js:33:14) # at Socket.g (events.js:291:16) # at emitNone (events.js:86:13) # at Socket.emit (events.js:185:7) # at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1075:10) --- duration_ms: 0.380 ``` /cc @mhdawson ?
process
investigate flaky parallel test child process fork regr gh version master platform aix subsystem child process this test failure recently occurred on aix output not ok parallel test child process fork regr gh home iojs build workspace node test commit aix nodes test parallel test child process fork regr gh js throw err error channel closed at childprocess target send internal child process js at worker send cluster js at socket home iojs build workspace node test commit aix nodes test parallel test child process fork regr gh js at socket g events js at emitnone events js at socket emit events js at tcpconnectwrap afterconnect net js duration ms cc mhdawson
1
2,880
5,842,024,982
IssuesEvent
2017-05-10 03:49:43
inasafe/inasafe
https://api.github.com/repos/inasafe/inasafe
closed
Update post_processor_additional_rice with value from layer's keywords
Needs review Postprocessing
### Problem Currently, the formula to calculate the additional rice is using hard code value for `pregnant_rate` and `breastfeeding_rate`. Since in v4.1 we introduce pregnant and lactating field, it will be great if we can use it to calculate the additional rice. For the first implementation, we will merge the rate of pregnancy and lactating. cc @timlinux @Charlotte-Morgan @samnawi
1.0
Update post_processor_additional_rice with value from layer's keywords - ### Problem Currently, the formula to calculate the additional rice is using hard code value for `pregnant_rate` and `breastfeeding_rate`. Since in v4.1 we introduce pregnant and lactating field, it will be great if we can use it to calculate the additional rice. For the first implementation, we will merge the rate of pregnancy and lactating. cc @timlinux @Charlotte-Morgan @samnawi
process
update post processor additional rice with value from layer s keywords problem currently the formula to calculate the additional rice is using hard code value for pregnant rate and breastfeeding rate since in we introduce pregnant and lactating field it will be great if we can use it to calculate the additional rice for the first implementation we will merge the rate of pregnancy and lactating cc timlinux charlotte morgan samnawi
1
3,895
6,821,024,222
IssuesEvent
2017-11-07 15:37:40
syndesisio/syndesis-ui
https://api.github.com/repos/syndesisio/syndesis-ui
closed
Update ng2-dynamic-forms dependency to ng-dynamic-forms
dev process Priority - High
Noticed this while poking at CI: ``` * * * * * * * * * * * * * * * * * * * * * * * * * * !!! PLEASE NOTE !!! * * * * @ng2-dynamic-forms has been renamed to: * * * * @ng-dynamic-forms * * * * * * Please update your dependencies * * * * * * * * * * * * * * * * * * * * * * * * * * ```
1.0
Update ng2-dynamic-forms dependency to ng-dynamic-forms - Noticed this while poking at CI: ``` * * * * * * * * * * * * * * * * * * * * * * * * * * !!! PLEASE NOTE !!! * * * * @ng2-dynamic-forms has been renamed to: * * * * @ng-dynamic-forms * * * * * * Please update your dependencies * * * * * * * * * * * * * * * * * * * * * * * * * * ```
process
update dynamic forms dependency to ng dynamic forms noticed this while poking at ci please note dynamic forms has been renamed to ng dynamic forms please update your dependencies
1
12,590
14,991,913,500
IssuesEvent
2021-01-29 09:06:39
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
closed
BE: System Status API
p1 story team:data processing
### Description Backend component that allows retrieve the current status of the system. ### Related Services Which backend services must change for this story to be completed? ### Designs Paste the link to your designs here ### Acceptance Criteria The API should return the status for: - Data Sources (CloudSec, Log Analysis, SaaS) - Data Parsing (Classification, Normalization) - Detections (Rules, Policies, Scheduled Queries) - Delivery to Data Lake Status (Snowflake/Athena) - Alert Delivery Status
1.0
BE: System Status API - ### Description Backend component that allows retrieve the current status of the system. ### Related Services Which backend services must change for this story to be completed? ### Designs Paste the link to your designs here ### Acceptance Criteria The API should return the status for: - Data Sources (CloudSec, Log Analysis, SaaS) - Data Parsing (Classification, Normalization) - Detections (Rules, Policies, Scheduled Queries) - Delivery to Data Lake Status (Snowflake/Athena) - Alert Delivery Status
process
be system status api description backend component that allows retrieve the current status of the system related services which backend services must change for this story to be completed designs paste the link to your designs here acceptance criteria the api should return the status for data sources cloudsec log analysis saas data parsing classification normalization detections rules policies scheduled queries delivery to data lake status snowflake athena alert delivery status
1
132,691
18,268,843,272
IssuesEvent
2021-10-04 11:42:05
artsking/linux-3.0.35
https://api.github.com/repos/artsking/linux-3.0.35
opened
CVE-2013-2893 (Medium) detected in multiple libraries
security vulnerability
## CVE-2013-2893 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv3.8.6</b>, <b>linux-stable-rtv3.8.6</b>, <b>linux-stable-rtv3.8.6</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Human Interface Device (HID) subsystem in the Linux kernel through 3.11, when CONFIG_LOGITECH_FF, CONFIG_LOGIG940_FF, or CONFIG_LOGIWHEELS_FF is enabled, allows physically proximate attackers to cause a denial of service (heap-based out-of-bounds write) via a crafted device, related to (1) drivers/hid/hid-lgff.c, (2) drivers/hid/hid-lg3ff.c, and (3) drivers/hid/hid-lg4ff.c. <p>Publish Date: 2013-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-2893>CVE-2013-2893</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-2893">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-2893</a></p> <p>Release Date: 2013-09-16</p> <p>Fix Resolution: v3.12-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2013-2893 (Medium) detected in multiple libraries - ## CVE-2013-2893 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv3.8.6</b>, <b>linux-stable-rtv3.8.6</b>, <b>linux-stable-rtv3.8.6</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Human Interface Device (HID) subsystem in the Linux kernel through 3.11, when CONFIG_LOGITECH_FF, CONFIG_LOGIG940_FF, or CONFIG_LOGIWHEELS_FF is enabled, allows physically proximate attackers to cause a denial of service (heap-based out-of-bounds write) via a crafted device, related to (1) drivers/hid/hid-lgff.c, (2) drivers/hid/hid-lg3ff.c, and (3) drivers/hid/hid-lg4ff.c. <p>Publish Date: 2013-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-2893>CVE-2013-2893</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-2893">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-2893</a></p> <p>Release Date: 2013-09-16</p> <p>Fix Resolution: v3.12-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries linux stable linux stable linux stable vulnerability details the human interface device hid subsystem in the linux kernel through when config logitech ff config ff or config logiwheels ff is enabled allows physically proximate attackers to cause a denial of service heap based out of bounds write via a crafted device related to drivers hid hid lgff c drivers hid hid c and drivers hid hid c publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
375,304
11,102,405,072
IssuesEvent
2019-12-16 23:56:47
mapbox/mapbox-events-android
https://api.github.com/repos/mapbox/mapbox-events-android
closed
🐞 Test App in Debug Mode: Network (OkHttp) request seems to fail in strict mode. This happens only in Android O) (26 and 27). Works well in Pre 26 versions.
bug medium priority size:5 sprint 1 sprint:2
Test App in Debug Mode: OKhttp request seems to fail in strict mode. This happens only in Android O) (26 and 27). Works well in Pre 26 versions. **Configuration** - Android OS Version: 9 - Device model(s): Google Pixel, Nexus e.t.c. Dependent on Android Version than device. - App: Events Test App **Steps to Reproduce** 1. Install Test app in android-events repository from Android studio. 2. Stays on blank screen for a while and eventually the app gets closed. **Expected** Should show Mapbox Telemetry test screen with respondent UI.. **Logging** 2019-08-30 10:57:36.524 7808-7855/? D/StrictMode: StrictMode policy violation: android.os.strictmode.UntaggedSocketViolation: Untagged socket detected; use TrafficStats.setThreadSocketTag() to track all network usage at android.os.StrictMode.onUntaggedSocket(StrictMode.java:2023) at com.android.server.NetworkManagementSocketTagger.tag(NetworkManagementSocketTagger.java:82) at libcore.io.BlockGuardOs.tagSocket(BlockGuardOs.java:50) at libcore.io.BlockGuardOs.socket(BlockGuardOs.java:313) at libcore.io.IoBridge.socket(IoBridge.java:640) at java.net.PlainSocketImpl.socketCreate(PlainSocketImpl.java:124) at java.net.AbstractPlainSocketImpl.create(AbstractPlainSocketImpl.java:128) at java.net.Socket.createImpl(Socket.java:489) at java.net.Socket.getImpl(Socket.java:552) at java.net.Socket.setSoTimeout(Socket.java:1180) at okhttp3.internal.connection.RealConnection.connectSocket(RealConnection.java:244) at okhttp3.internal.connection.RealConnection.connect(RealConnection.java:166) at okhttp3.internal.connection.StreamAllocation.findConnection(StreamAllocation.java:257) at okhttp3.internal.connection.StreamAllocation.findHealthyConnection(StreamAllocation.java:135) at okhttp3.internal.connection.StreamAllocation.newStream(StreamAllocation.java:114) at okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.java:42) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121) at okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.java:93) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121) at okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.java:93) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.java:126) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121) at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:254) at okhttp3.RealCall$AsyncCall.execute(RealCall.java:200) at okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1167) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:641) at java.lang.Thread.run(Thread.java:764)
1.0
🐞 Test App in Debug Mode: Network (OkHttp) request seems to fail in strict mode. This happens only in Android O) (26 and 27). Works well in Pre 26 versions. - Test App in Debug Mode: OKhttp request seems to fail in strict mode. This happens only in Android O) (26 and 27). Works well in Pre 26 versions. **Configuration** - Android OS Version: 9 - Device model(s): Google Pixel, Nexus e.t.c. Dependent on Android Version than device. - App: Events Test App **Steps to Reproduce** 1. Install Test app in android-events repository from Android studio. 2. Stays on blank screen for a while and eventually the app gets closed. **Expected** Should show Mapbox Telemetry test screen with respondent UI.. **Logging** 2019-08-30 10:57:36.524 7808-7855/? D/StrictMode: StrictMode policy violation: android.os.strictmode.UntaggedSocketViolation: Untagged socket detected; use TrafficStats.setThreadSocketTag() to track all network usage at android.os.StrictMode.onUntaggedSocket(StrictMode.java:2023) at com.android.server.NetworkManagementSocketTagger.tag(NetworkManagementSocketTagger.java:82) at libcore.io.BlockGuardOs.tagSocket(BlockGuardOs.java:50) at libcore.io.BlockGuardOs.socket(BlockGuardOs.java:313) at libcore.io.IoBridge.socket(IoBridge.java:640) at java.net.PlainSocketImpl.socketCreate(PlainSocketImpl.java:124) at java.net.AbstractPlainSocketImpl.create(AbstractPlainSocketImpl.java:128) at java.net.Socket.createImpl(Socket.java:489) at java.net.Socket.getImpl(Socket.java:552) at java.net.Socket.setSoTimeout(Socket.java:1180) at okhttp3.internal.connection.RealConnection.connectSocket(RealConnection.java:244) at okhttp3.internal.connection.RealConnection.connect(RealConnection.java:166) at okhttp3.internal.connection.StreamAllocation.findConnection(StreamAllocation.java:257) at okhttp3.internal.connection.StreamAllocation.findHealthyConnection(StreamAllocation.java:135) at okhttp3.internal.connection.StreamAllocation.newStream(StreamAllocation.java:114) at okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.java:42) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121) at okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.java:93) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121) at okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.java:93) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.java:126) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:147) at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.java:121) at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:254) at okhttp3.RealCall$AsyncCall.execute(RealCall.java:200) at okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1167) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:641) at java.lang.Thread.run(Thread.java:764)
non_process
🐞 test app in debug mode network okhttp request seems to fail in strict mode this happens only in android o and works well in pre versions test app in debug mode okhttp request seems to fail in strict mode this happens only in android o and works well in pre versions configuration android os version device model s google pixel nexus e t c dependent on android version than device app events test app steps to reproduce install test app in android events repository from android studio stays on blank screen for a while and eventually the app gets closed expected should show mapbox telemetry test screen with respondent ui logging d strictmode strictmode policy violation android os strictmode untaggedsocketviolation untagged socket detected use trafficstats setthreadsockettag to track all network usage at android os strictmode onuntaggedsocket strictmode java at com android server networkmanagementsockettagger tag networkmanagementsockettagger java at libcore io blockguardos tagsocket blockguardos java at libcore io blockguardos socket blockguardos java at libcore io iobridge socket iobridge java at java net plainsocketimpl socketcreate plainsocketimpl java at java net abstractplainsocketimpl create abstractplainsocketimpl java at java net socket createimpl socket java at java net socket getimpl socket java at java net socket setsotimeout socket java at internal connection realconnection connectsocket realconnection java at internal connection realconnection connect realconnection java at internal connection streamallocation findconnection streamallocation java at internal connection streamallocation findhealthyconnection streamallocation java at internal connection streamallocation newstream streamallocation java at internal connection connectinterceptor intercept connectinterceptor java at internal http realinterceptorchain proceed realinterceptorchain java at internal http realinterceptorchain proceed realinterceptorchain java at internal cache cacheinterceptor intercept cacheinterceptor java at internal http realinterceptorchain proceed realinterceptorchain java at internal http realinterceptorchain proceed realinterceptorchain java at internal http bridgeinterceptor intercept bridgeinterceptor java at internal http realinterceptorchain proceed realinterceptorchain java at internal http retryandfollowupinterceptor intercept retryandfollowupinterceptor java at internal http realinterceptorchain proceed realinterceptorchain java at internal http realinterceptorchain proceed realinterceptorchain java at realcall getresponsewithinterceptorchain realcall java at realcall asynccall execute realcall java at internal namedrunnable run namedrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
0
18,721
24,611,127,575
IssuesEvent
2022-10-14 21:37:21
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Release checklist 0.66
enhancement process
### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.62.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [x] Deploy to VM ## Performance - [x] Deploy to Kubernetes - [x] Deploy to VM - [x] gRPC API performance tests - [x] Importer performance tests - [x] REST API performance tests - [x] Migrations tested against mainnet clone ## Previewnet - [x] Deploy to VM ## Staging - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA ## Testnet - [ ] Deploy to VM ## Mainnet - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA - [ ] Deploy to VM - [ ] Deploy to ETL ### Alternatives _No response_
1.0
Release checklist 0.66 - ### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.62.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [x] Deploy to VM ## Performance - [x] Deploy to Kubernetes - [x] Deploy to VM - [x] gRPC API performance tests - [x] Importer performance tests - [x] REST API performance tests - [x] Migrations tested against mainnet clone ## Previewnet - [x] Deploy to VM ## Staging - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA ## Testnet - [ ] Deploy to VM ## Mainnet - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA - [ ] Deploy to VM - [ ] Deploy to ETL ### Alternatives _No response_
process
release checklist problem we need a checklist to verify the release is rolled out successfully solution preparation milestone field populated on relevant nothing open for github checks for branch are passing automated kubernetes deployment successful tag release upload release artifacts publish release integration deploy to vm performance deploy to kubernetes deploy to vm grpc api performance tests importer performance tests rest api performance tests migrations tested against mainnet clone previewnet deploy to vm staging deploy to kubernetes eu deploy to kubernetes na testnet deploy to vm mainnet deploy to kubernetes eu deploy to kubernetes na deploy to vm deploy to etl alternatives no response
1
7,069
10,220,552,520
IssuesEvent
2019-08-15 21:42:11
Jeffail/benthos
https://api.github.com/repos/Jeffail/benthos
closed
Text/prepend operator misbehaves
bug processors
I found this peculiar error where such simplest pipeline fails to deliver proper output. First build anything that uses `text` processor with `prepend` operator, eg.: ``` input: type: benthos_lab pipeline: processors: - type: text text: operator: prepend value: "--" output: type: benthos_lab ``` Then prepare input; remember to keep it under 6 bytes (as total lenght with prefix should not exceed 8 bytes), eg.: ``` aaaaaa bbbb cc ``` Observe output: ``` Compiled successfully. --ccbbaa --ccbb --cc ``` Where proper output should be: ``` --aaaaaa --bbbb --cc ``` (Lab example can be found here: https://lab.benthos.dev/l/6htCxZXE7z2 )
1.0
Text/prepend operator misbehaves - I found this peculiar error where such simplest pipeline fails to deliver proper output. First build anything that uses `text` processor with `prepend` operator, eg.: ``` input: type: benthos_lab pipeline: processors: - type: text text: operator: prepend value: "--" output: type: benthos_lab ``` Then prepare input; remember to keep it under 6 bytes (as total lenght with prefix should not exceed 8 bytes), eg.: ``` aaaaaa bbbb cc ``` Observe output: ``` Compiled successfully. --ccbbaa --ccbb --cc ``` Where proper output should be: ``` --aaaaaa --bbbb --cc ``` (Lab example can be found here: https://lab.benthos.dev/l/6htCxZXE7z2 )
process
text prepend operator misbehaves i found this peculiar error where such simplest pipeline fails to deliver proper output first build anything that uses text processor with prepend operator eg input type benthos lab pipeline processors type text text operator prepend value output type benthos lab then prepare input remember to keep it under bytes as total lenght with prefix should not exceed bytes eg aaaaaa bbbb cc observe output compiled successfully ccbbaa ccbb cc where proper output should be aaaaaa bbbb cc lab example can be found here
1
18,711
24,603,783,650
IssuesEvent
2022-10-14 14:32:18
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[DID] [Discard FHIR after DID Enabled] Patient resource > Value of the 'active' key is not updating to the 'false' when participant is withdrawn from the study or deactivates his/her account
Bug P1 Response datastore Process: Fixed Process: Tested dev
AR: Patient resource > Value of the 'active' key is not updating to the 'false' when the participant is withdrawn from the study or deactivates his/her account ER: The value of the 'active' key should be changed from true to false when the participant withdraws from the study ![image](https://user-images.githubusercontent.com/71445210/186413354-8701167f-5fa3-42f4-ae25-53a1a10c6d0e.png)
2.0
[DID] [Discard FHIR after DID Enabled] Patient resource > Value of the 'active' key is not updating to the 'false' when participant is withdrawn from the study or deactivates his/her account - AR: Patient resource > Value of the 'active' key is not updating to the 'false' when the participant is withdrawn from the study or deactivates his/her account ER: The value of the 'active' key should be changed from true to false when the participant withdraws from the study ![image](https://user-images.githubusercontent.com/71445210/186413354-8701167f-5fa3-42f4-ae25-53a1a10c6d0e.png)
process
patient resource value of the active key is not updating to the false when participant is withdrawn from the study or deactivates his her account ar patient resource value of the active key is not updating to the false when the participant is withdrawn from the study or deactivates his her account er the value of the active key should be changed from true to false when the participant withdraws from the study
1
739,911
25,728,040,527
IssuesEvent
2022-12-07 18:02:11
slynch8/10x
https://api.github.com/repos/slynch8/10x
closed
Feature request: Notepad++ like temp files
feature Priority 3 current
Let explain this feature: - I am able to create a new file in notepad++ - I can freely edit it at will (with undo etc) - I don't need to save it - When I exit notepad++ the file is stored somewhere in its installation folders or temp folders - When I restart notepad++ it restores the state of opened tabs, including unsaved files - This also propagates to project files, so if I edit a .c file from a project and never save it, when I restart notepad++ it will remember that the file has been changed and also load the changes. So in a nutshell notepad++ remembers changes in both saved and unsaved files across sessions. If this is tricky to do in 10x, I would at least appreciate if 10xnotepad could have this ability.
1.0
Feature request: Notepad++ like temp files - Let explain this feature: - I am able to create a new file in notepad++ - I can freely edit it at will (with undo etc) - I don't need to save it - When I exit notepad++ the file is stored somewhere in its installation folders or temp folders - When I restart notepad++ it restores the state of opened tabs, including unsaved files - This also propagates to project files, so if I edit a .c file from a project and never save it, when I restart notepad++ it will remember that the file has been changed and also load the changes. So in a nutshell notepad++ remembers changes in both saved and unsaved files across sessions. If this is tricky to do in 10x, I would at least appreciate if 10xnotepad could have this ability.
non_process
feature request notepad like temp files let explain this feature i am able to create a new file in notepad i can freely edit it at will with undo etc i don t need to save it when i exit notepad the file is stored somewhere in its installation folders or temp folders when i restart notepad it restores the state of opened tabs including unsaved files this also propagates to project files so if i edit a c file from a project and never save it when i restart notepad it will remember that the file has been changed and also load the changes so in a nutshell notepad remembers changes in both saved and unsaved files across sessions if this is tricky to do in i would at least appreciate if could have this ability
0
11,075
13,911,274,369
IssuesEvent
2020-10-20 17:09:53
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
opened
Have a checker that inspects auxiliaty_data_dir and exit if desired data is not there
enhancement preprocessor
**Is your feature request related to a problem? Please describe.** Not really a problem, but if you run a recipe as n00b (like I did with the Carvalhais one) and you don't bother check the data requirements, you'll bite the bait and run the whole show only to have the thing die on you at diagnostic stage coz some aux data ain't there. I can see this happen to a lot of first timers, and some of them will not even bother reading the recipe documentation and straight open an issue saying the thing's broken. **Would you be able to help out?** Yeah, for sure! I'm thinking the bext way is to have it part of the data finder moduleand initialized from the recipevia an additional field eg `auxiliary_data`
1.0
Have a checker that inspects auxiliaty_data_dir and exit if desired data is not there - **Is your feature request related to a problem? Please describe.** Not really a problem, but if you run a recipe as n00b (like I did with the Carvalhais one) and you don't bother check the data requirements, you'll bite the bait and run the whole show only to have the thing die on you at diagnostic stage coz some aux data ain't there. I can see this happen to a lot of first timers, and some of them will not even bother reading the recipe documentation and straight open an issue saying the thing's broken. **Would you be able to help out?** Yeah, for sure! I'm thinking the bext way is to have it part of the data finder moduleand initialized from the recipevia an additional field eg `auxiliary_data`
process
have a checker that inspects auxiliaty data dir and exit if desired data is not there is your feature request related to a problem please describe not really a problem but if you run a recipe as like i did with the carvalhais one and you don t bother check the data requirements you ll bite the bait and run the whole show only to have the thing die on you at diagnostic stage coz some aux data ain t there i can see this happen to a lot of first timers and some of them will not even bother reading the recipe documentation and straight open an issue saying the thing s broken would you be able to help out yeah for sure i m thinking the bext way is to have it part of the data finder moduleand initialized from the recipevia an additional field eg auxiliary data
1
44,376
9,575,124,551
IssuesEvent
2019-05-07 05:05:19
happycube/ld-decode
https://api.github.com/repos/happycube/ld-decode
closed
regression: NTSC disks with off-spec color burst phase are failing phase ID detection
bug ld-decode
Simon found it using: python3 ld-decode.py /media/sdi/nas/SFTP/LDV4300D_2\ Captures/National\ Gallery\ of\ Art/National\Gallery\ of\ Art_CAV_NTSC_side1_2018-12-12_14-57-06.lds ~/Capture/ngoa --start 12000 --length 200 Note that this disk is well made, aside from the off-center color burst. So a sample of this should be kept in ld-decode-testdata. The fix here will be to have a "sticky" indication on if NTSC phase detection needs to be rotated by 90 degrees in a vaguely clean way.
1.0
regression: NTSC disks with off-spec color burst phase are failing phase ID detection - Simon found it using: python3 ld-decode.py /media/sdi/nas/SFTP/LDV4300D_2\ Captures/National\ Gallery\ of\ Art/National\Gallery\ of\ Art_CAV_NTSC_side1_2018-12-12_14-57-06.lds ~/Capture/ngoa --start 12000 --length 200 Note that this disk is well made, aside from the off-center color burst. So a sample of this should be kept in ld-decode-testdata. The fix here will be to have a "sticky" indication on if NTSC phase detection needs to be rotated by 90 degrees in a vaguely clean way.
non_process
regression ntsc disks with off spec color burst phase are failing phase id detection simon found it using ld decode py media sdi nas sftp captures national gallery of art national gallery of art cav ntsc lds capture ngoa start length note that this disk is well made aside from the off center color burst so a sample of this should be kept in ld decode testdata the fix here will be to have a sticky indication on if ntsc phase detection needs to be rotated by degrees in a vaguely clean way
0
237,678
19,665,408,225
IssuesEvent
2022-01-10 21:51:27
surveysolutions/surveysolutions
https://api.github.com/repos/surveysolutions/surveysolutions
closed
Allow questionnaires with questions in the Cover page only
enhancement testing required
As Designer user I want to have an ability to create a questionnaire with questions in the Cover section only so that I will be able to preload answers for my interview in the easiest way via uploading or manually. The ticket can be labeled as Verified when: 1. Questionnaires without questions in the First section (not Cover) can be compiled. 2. Web-tester, Tester app, IN app, SV app, Web-interviewer are works in appropriate way with such questionnaires.
1.0
Allow questionnaires with questions in the Cover page only - As Designer user I want to have an ability to create a questionnaire with questions in the Cover section only so that I will be able to preload answers for my interview in the easiest way via uploading or manually. The ticket can be labeled as Verified when: 1. Questionnaires without questions in the First section (not Cover) can be compiled. 2. Web-tester, Tester app, IN app, SV app, Web-interviewer are works in appropriate way with such questionnaires.
non_process
allow questionnaires with questions in the cover page only as designer user i want to have an ability to create a questionnaire with questions in the cover section only so that i will be able to preload answers for my interview in the easiest way via uploading or manually the ticket can be labeled as verified when questionnaires without questions in the first section not cover can be compiled web tester tester app in app sv app web interviewer are works in appropriate way with such questionnaires
0
1,765
3,367,916,889
IssuesEvent
2015-11-22 15:53:40
eslint/eslint
https://api.github.com/repos/eslint/eslint
opened
Shrink wrap dependencies?
infrastructure proposal
We just had an issue (#4481) where a dependency patch upgrade introduced a problem for some users. I'm wondering if it makes sense to shrink wrap our dependencies so this doesn't happen in the future? The advantage of doing so is that we can know with 100% certainty which versions of which dependencies people are using. The disadvantage is that we will need to pay close attention to patch and minor releases of our dependencies. With each minor release, we would need to manually upgrade dependencies. Thoughts?
1.0
Shrink wrap dependencies? - We just had an issue (#4481) where a dependency patch upgrade introduced a problem for some users. I'm wondering if it makes sense to shrink wrap our dependencies so this doesn't happen in the future? The advantage of doing so is that we can know with 100% certainty which versions of which dependencies people are using. The disadvantage is that we will need to pay close attention to patch and minor releases of our dependencies. With each minor release, we would need to manually upgrade dependencies. Thoughts?
non_process
shrink wrap dependencies we just had an issue where a dependency patch upgrade introduced a problem for some users i m wondering if it makes sense to shrink wrap our dependencies so this doesn t happen in the future the advantage of doing so is that we can know with certainty which versions of which dependencies people are using the disadvantage is that we will need to pay close attention to patch and minor releases of our dependencies with each minor release we would need to manually upgrade dependencies thoughts
0
77,071
15,496,267,469
IssuesEvent
2021-03-11 02:21:46
n-devs/share-we-go
https://api.github.com/repos/n-devs/share-we-go
opened
CVE-2020-15366 (Medium) detected in ajv-6.10.0.tgz
security vulnerability
## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ajv-6.10.0.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-6.10.0.tgz">https://registry.npmjs.org/ajv/-/ajv-6.10.0.tgz</a></p> <p>Path to dependency file: /share-we-go/package.json</p> <p>Path to vulnerable library: share-we-go/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.0.0.tgz (Root Library) - eslint-5.16.0.tgz - :x: **ajv-6.10.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: ajv - 6.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15366 (Medium) detected in ajv-6.10.0.tgz - ## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ajv-6.10.0.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-6.10.0.tgz">https://registry.npmjs.org/ajv/-/ajv-6.10.0.tgz</a></p> <p>Path to dependency file: /share-we-go/package.json</p> <p>Path to vulnerable library: share-we-go/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.0.0.tgz (Root Library) - eslint-5.16.0.tgz - :x: **ajv-6.10.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: ajv - 6.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in ajv tgz cve medium severity vulnerability vulnerable library ajv tgz another json schema validator library home page a href path to dependency file share we go package json path to vulnerable library share we go node modules ajv package json dependency hierarchy react scripts tgz root library eslint tgz x ajv tgz vulnerable library vulnerability details an issue was discovered in ajv validate in ajv aka another json schema validator a carefully crafted json schema could be provided that allows execution of other code by prototype pollution while untrusted schemas are recommended against the worst case of an untrusted schema should be a denial of service not execution of code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ajv step up your open source security game with whitesource
0
99,571
8,704,609,994
IssuesEvent
2018-12-05 19:53:31
chartjs/Chart.js
https://api.github.com/repos/chartjs/Chart.js
closed
Chart Element Shows Two tooltips
status: needs test case
Hey Guys when I hover over my chart, it is displaying two tooltips at the same time. Is there a way around this behavior ? Isn’t it more logical to only display one tooltip at a time (the one relating to the point which is closest ? )… Currently the tooltips are overlapping and it gets quite frustrating for someone trying to read the chart. The double tooltips continue to happen unless the points are very spread out from each other:
1.0
Chart Element Shows Two tooltips - Hey Guys when I hover over my chart, it is displaying two tooltips at the same time. Is there a way around this behavior ? Isn’t it more logical to only display one tooltip at a time (the one relating to the point which is closest ? )… Currently the tooltips are overlapping and it gets quite frustrating for someone trying to read the chart. The double tooltips continue to happen unless the points are very spread out from each other:
non_process
chart element shows two tooltips hey guys when i hover over my chart it is displaying two tooltips at the same time is there a way around this behavior isn’t it more logical to only display one tooltip at a time the one relating to the point which is closest … currently the tooltips are overlapping and it gets quite frustrating for someone trying to read the chart the double tooltips continue to happen unless the points are very spread out from each other
0
6,818
9,961,757,731
IssuesEvent
2019-07-07 08:20:07
linnovate/root
https://api.github.com/repos/linnovate/root
closed
Can't reset profile picture
2.0.7 Fixed Process bug Profile bug
open the profile page. click on replace. pick some picture. click on reset. the picture still stay as the one you replace. ![image](https://user-images.githubusercontent.com/31100069/50400311-6f907780-078e-11e9-9f9b-501dd292811e.png)
1.0
Can't reset profile picture - open the profile page. click on replace. pick some picture. click on reset. the picture still stay as the one you replace. ![image](https://user-images.githubusercontent.com/31100069/50400311-6f907780-078e-11e9-9f9b-501dd292811e.png)
process
can t reset profile picture open the profile page click on replace pick some picture click on reset the picture still stay as the one you replace
1
6,389
9,462,639,248
IssuesEvent
2019-04-17 15:50:44
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Updates to Submitted application page
Apply Process Requirements Ready State Dept.
Who: Student What: Update/Withdraw page Why: As a student I would like the ability to update or withdraw my application and need an easy way to accomplish this Acceptance Criteria: - From the landing page Applied section, when the intern clicks on the link in the Program column, they will be taken to the Submitted application page where they see the Submitted application section - The header reads, "Submitted application" - The text under the header reads, "You have submitted the following application to the <community name> for <cycle name>. - There will be two buttons: Update and Withdraw - Update will be handled with ticket 3231 - Withdraw is handled with ticket 3087 - Under the update and withdraw buttons, the header reads, "Applying to these internship opportunities" <NOTE question for Alex on this title> - Then their first, second, and third choices will display Questions: - Will they still get the update modal? - Julie - They will get the modal if they click the "update" button. We need to make sure they understand they have to resubmit their application not just make edits. - Will they also go to this page when they click Update application from the internship? - They would likely also go to this page if they click Apply from an internship they didn't select but are clicking apply on after they submitted their 3 - What about an in process application? They would skip this page if they go back to complete correct? - Julie- I think they would skip this page because they are not updating an internship they are completing an application in process. Invision Mock: https://opm.invisionapp.com/share/ZEPNZR09Q54#/352514657_State_-_View_Selections_-_Update_Or_Withdraw_-Desktop-
1.0
Updates to Submitted application page - Who: Student What: Update/Withdraw page Why: As a student I would like the ability to update or withdraw my application and need an easy way to accomplish this Acceptance Criteria: - From the landing page Applied section, when the intern clicks on the link in the Program column, they will be taken to the Submitted application page where they see the Submitted application section - The header reads, "Submitted application" - The text under the header reads, "You have submitted the following application to the <community name> for <cycle name>. - There will be two buttons: Update and Withdraw - Update will be handled with ticket 3231 - Withdraw is handled with ticket 3087 - Under the update and withdraw buttons, the header reads, "Applying to these internship opportunities" <NOTE question for Alex on this title> - Then their first, second, and third choices will display Questions: - Will they still get the update modal? - Julie - They will get the modal if they click the "update" button. We need to make sure they understand they have to resubmit their application not just make edits. - Will they also go to this page when they click Update application from the internship? - They would likely also go to this page if they click Apply from an internship they didn't select but are clicking apply on after they submitted their 3 - What about an in process application? They would skip this page if they go back to complete correct? - Julie- I think they would skip this page because they are not updating an internship they are completing an application in process. Invision Mock: https://opm.invisionapp.com/share/ZEPNZR09Q54#/352514657_State_-_View_Selections_-_Update_Or_Withdraw_-Desktop-
process
updates to submitted application page who student what update withdraw page why as a student i would like the ability to update or withdraw my application and need an easy way to accomplish this acceptance criteria from the landing page applied section when the intern clicks on the link in the program column they will be taken to the submitted application page where they see the submitted application section the header reads submitted application the text under the header reads you have submitted the following application to the for there will be two buttons update and withdraw update will be handled with ticket withdraw is handled with ticket under the update and withdraw buttons the header reads applying to these internship opportunities then their first second and third choices will display questions will they still get the update modal julie they will get the modal if they click the update button we need to make sure they understand they have to resubmit their application not just make edits will they also go to this page when they click update application from the internship they would likely also go to this page if they click apply from an internship they didn t select but are clicking apply on after they submitted their what about an in process application they would skip this page if they go back to complete correct julie i think they would skip this page because they are not updating an internship they are completing an application in process invision mock
1
23,297
11,904,225,428
IssuesEvent
2020-03-30 16:30:33
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
Regression in flutter_gallery_ios32__transition_perf benchmarks
engine perf: speed severe: performance t: benchmark ⚠ TODAY
`90th_percentile_frame_build_time_millis`, `90th_percentile_frame_rasterizer_time_millis`, `99th_percentile_frame_rasterizer_time_millis`. Very clear signal: ![Screen Shot 2020-03-30 at 9 29 22 AM](https://user-images.githubusercontent.com/1435716/77937283-f6bd7c80-7268-11ea-860c-5f0885ddbc74.png) Looks like this engine roll: https://github.com/flutter/flutter/pull/53403
True
Regression in flutter_gallery_ios32__transition_perf benchmarks - `90th_percentile_frame_build_time_millis`, `90th_percentile_frame_rasterizer_time_millis`, `99th_percentile_frame_rasterizer_time_millis`. Very clear signal: ![Screen Shot 2020-03-30 at 9 29 22 AM](https://user-images.githubusercontent.com/1435716/77937283-f6bd7c80-7268-11ea-860c-5f0885ddbc74.png) Looks like this engine roll: https://github.com/flutter/flutter/pull/53403
non_process
regression in flutter gallery transition perf benchmarks percentile frame build time millis percentile frame rasterizer time millis percentile frame rasterizer time millis very clear signal looks like this engine roll
0
128,267
27,223,937,347
IssuesEvent
2023-02-21 08:19:14
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] Backend article panel
No Code Attached Yet J4 Backend Template
Dear developers! Since there has recently been a [debate](https://github.com/joomla-projects/j4adminui/issues/186) about the admin panel of Atum or Khonsu. And after a long argument, the Honxu developers closed the topic until next time. But there were people who said: that you can implement the best sides of Khonsu in Atum. And for this I suggest viewing the topic: Creation of material on Joomla4: **This is how it looks in Atum:** ![Atum](https://user-images.githubusercontent.com/40716604/75892334-f4772680-5e52-11ea-88d7-54c8be2d92a9.png) **And so in Khonsu:** ![Khonsu](https://user-images.githubusercontent.com/40716604/75892370-01941580-5e53-11ea-9394-c3831d9ea557.png) I think the result is on the face. In fulfillment of Khonsu, I think it looks well designed and modern. All the necessary tabs are rendered to the right and convenient for filling (tabs). And if this is implemented, then I think not much to change the Atum code. And that would get a good response from the community. What do you think? can this be realized in Atum? @angieradtke I am writing through google translate
1.0
[4.0] Backend article panel - Dear developers! Since there has recently been a [debate](https://github.com/joomla-projects/j4adminui/issues/186) about the admin panel of Atum or Khonsu. And after a long argument, the Honxu developers closed the topic until next time. But there were people who said: that you can implement the best sides of Khonsu in Atum. And for this I suggest viewing the topic: Creation of material on Joomla4: **This is how it looks in Atum:** ![Atum](https://user-images.githubusercontent.com/40716604/75892334-f4772680-5e52-11ea-88d7-54c8be2d92a9.png) **And so in Khonsu:** ![Khonsu](https://user-images.githubusercontent.com/40716604/75892370-01941580-5e53-11ea-9394-c3831d9ea557.png) I think the result is on the face. In fulfillment of Khonsu, I think it looks well designed and modern. All the necessary tabs are rendered to the right and convenient for filling (tabs). And if this is implemented, then I think not much to change the Atum code. And that would get a good response from the community. What do you think? can this be realized in Atum? @angieradtke I am writing through google translate
non_process
backend article panel dear developers since there has recently been a about the admin panel of atum or khonsu and after a long argument the honxu developers closed the topic until next time but there were people who said that you can implement the best sides of khonsu in atum and for this i suggest viewing the topic creation of material on this is how it looks in atum and so in khonsu i think the result is on the face in fulfillment of khonsu i think it looks well designed and modern all the necessary tabs are rendered to the right and convenient for filling tabs and if this is implemented then i think not much to change the atum code and that would get a good response from the community what do you think can this be realized in atum angieradtke i am writing through google translate
0
350,059
31,849,375,716
IssuesEvent
2023-09-14 23:20:18
hpc/charliecloud
https://api.github.com/repos/hpc/charliecloud
closed
xattrs tests fail if `/tmp` is a tmpfs
bug high test
If `/tmp` is a tmpfs (or some other filesystem that does not support xattrs), then various tests fail, for example: ``` ✗ cache: restore ACLs, xattrs [...] 5. RUN setfattr -n user.foo -v bar /home/foo setfattr: /home/foo: Not supported something went wrong, rolling back ... HEAD is now at 9e9809a RUN touch /home/foo error: build failed: RUN command exited with 1 trace: build.py:1180:execute, build.py:250:__default__, visitors.py:241:_call_userfunc, visitors.py:269:visit_topdown, build.py:186:main, ch-image:314:main ``` Fortunately, `/tmp` is not a tmpfs on GitHub Actions, so CI still passes. I don't think this necessarily needs to be fixed for 0.34. I think the thing to do is simply identify the problem tests and skip them if *both* (1) `/tmp` does not support xattrs and (2) we're not on GitHub Actions (i.e., `$GITHUB_ACTIONS` is not set).
1.0
xattrs tests fail if `/tmp` is a tmpfs - If `/tmp` is a tmpfs (or some other filesystem that does not support xattrs), then various tests fail, for example: ``` ✗ cache: restore ACLs, xattrs [...] 5. RUN setfattr -n user.foo -v bar /home/foo setfattr: /home/foo: Not supported something went wrong, rolling back ... HEAD is now at 9e9809a RUN touch /home/foo error: build failed: RUN command exited with 1 trace: build.py:1180:execute, build.py:250:__default__, visitors.py:241:_call_userfunc, visitors.py:269:visit_topdown, build.py:186:main, ch-image:314:main ``` Fortunately, `/tmp` is not a tmpfs on GitHub Actions, so CI still passes. I don't think this necessarily needs to be fixed for 0.34. I think the thing to do is simply identify the problem tests and skip them if *both* (1) `/tmp` does not support xattrs and (2) we're not on GitHub Actions (i.e., `$GITHUB_ACTIONS` is not set).
non_process
xattrs tests fail if tmp is a tmpfs if tmp is a tmpfs or some other filesystem that does not support xattrs then various tests fail for example ✗ cache restore acls xattrs run setfattr n user foo v bar home foo setfattr home foo not supported something went wrong rolling back head is now at run touch home foo error build failed run command exited with trace build py execute build py default visitors py call userfunc visitors py visit topdown build py main ch image main fortunately tmp is not a tmpfs on github actions so ci still passes i don t think this necessarily needs to be fixed for i think the thing to do is simply identify the problem tests and skip them if both tmp does not support xattrs and we re not on github actions i e github actions is not set
0
13,175
15,597,159,041
IssuesEvent
2021-03-18 16:37:06
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Add test for custom engine location from the Prisma Client JS instance
process/candidate team/client tech/typescript topic: binary topic: internal topic: prisma-client topic: tests
## Problem This internal feature can be used to point to a custom engine location from the Prisma Client JS instance: ```ts const prisma = new PrismaClient({ __internal: { engine: { binaryPath: path.join(__dirname, 'query-engine-debian-openssl-1.1.x') } } } as any) ``` Currently not covered by tests. ## Suggested solution Add test like ```ts import { PrismaClient } from '@prisma/client' import path from 'path' async function main() { const prisma = new PrismaClient({ __internal: { engine: { binaryPath: path.join(__dirname, 'query-engine-debian-openssl-1.1.x') } } } as any) const data = await prisma.user.findMany() console.log(data) const anyPrisma = prisma as any console.log(anyPrisma._engine.prismaPath) console.log(await anyPrisma._engine.getPrismaPath()) prisma.$disconnect() } main() ``` ## Alternatives Add a test in e2e-tests ## Additional context Issue on the internal feature and usage https://github.com/prisma/prisma/issues/3499 Related https://github.com/prisma/prisma/issues/6168
1.0
Add test for custom engine location from the Prisma Client JS instance - ## Problem This internal feature can be used to point to a custom engine location from the Prisma Client JS instance: ```ts const prisma = new PrismaClient({ __internal: { engine: { binaryPath: path.join(__dirname, 'query-engine-debian-openssl-1.1.x') } } } as any) ``` Currently not covered by tests. ## Suggested solution Add test like ```ts import { PrismaClient } from '@prisma/client' import path from 'path' async function main() { const prisma = new PrismaClient({ __internal: { engine: { binaryPath: path.join(__dirname, 'query-engine-debian-openssl-1.1.x') } } } as any) const data = await prisma.user.findMany() console.log(data) const anyPrisma = prisma as any console.log(anyPrisma._engine.prismaPath) console.log(await anyPrisma._engine.getPrismaPath()) prisma.$disconnect() } main() ``` ## Alternatives Add a test in e2e-tests ## Additional context Issue on the internal feature and usage https://github.com/prisma/prisma/issues/3499 Related https://github.com/prisma/prisma/issues/6168
process
add test for custom engine location from the prisma client js instance problem this internal feature can be used to point to a custom engine location from the prisma client js instance ts const prisma new prismaclient internal engine binarypath path join dirname query engine debian openssl x as any currently not covered by tests suggested solution add test like ts import prismaclient from prisma client import path from path async function main const prisma new prismaclient internal engine binarypath path join dirname query engine debian openssl x as any const data await prisma user findmany console log data const anyprisma prisma as any console log anyprisma engine prismapath console log await anyprisma engine getprismapath prisma disconnect main alternatives add a test in tests additional context issue on the internal feature and usage related
1
65,915
7,930,660,622
IssuesEvent
2018-07-06 19:51:17
skycoin/skycoin
https://api.github.com/repos/skycoin/skycoin
closed
UI Improvements
ui-design wallet
Several of the improvements made to the UI of the web wallet should be applied to the desktop wallet. The PRs in which the improvements are: - https://github.com/skycoin/skycoin-web/pull/326 - https://github.com/skycoin/skycoin-web/pull/391 (When removing the pipe it will be necessary to change the text on the blockchain page, something that was not done in the web wallet) - https://github.com/skycoin/skycoin-web/pull/384 - https://github.com/skycoin/skycoin-web/pull/374 - https://github.com/skycoin/skycoin-web/pull/371 - https://github.com/skycoin/skycoin-web/pull/318 (In this case it is only necessary to use the new version of `minus-red.png`)
1.0
UI Improvements - Several of the improvements made to the UI of the web wallet should be applied to the desktop wallet. The PRs in which the improvements are: - https://github.com/skycoin/skycoin-web/pull/326 - https://github.com/skycoin/skycoin-web/pull/391 (When removing the pipe it will be necessary to change the text on the blockchain page, something that was not done in the web wallet) - https://github.com/skycoin/skycoin-web/pull/384 - https://github.com/skycoin/skycoin-web/pull/374 - https://github.com/skycoin/skycoin-web/pull/371 - https://github.com/skycoin/skycoin-web/pull/318 (In this case it is only necessary to use the new version of `minus-red.png`)
non_process
ui improvements several of the improvements made to the ui of the web wallet should be applied to the desktop wallet the prs in which the improvements are when removing the pipe it will be necessary to change the text on the blockchain page something that was not done in the web wallet in this case it is only necessary to use the new version of minus red png
0
2,151
4,999,038,994
IssuesEvent
2016-12-09 21:52:56
gwpy/gwpy
https://api.github.com/repos/gwpy/gwpy
closed
resample() phase shift
needs-clarification signal-processing timeseries
Duncan The method gwpy.timeseries.StateVector.resample would be better if it ensured zero-phase shift between input and output. Roy
1.0
resample() phase shift - Duncan The method gwpy.timeseries.StateVector.resample would be better if it ensured zero-phase shift between input and output. Roy
process
resample phase shift duncan the method gwpy timeseries statevector resample would be better if it ensured zero phase shift between input and output roy
1
10,691
12,641,987,234
IssuesEvent
2020-06-16 07:21:42
multitheftauto/amx
https://api.github.com/repos/multitheftauto/amx
opened
EnablePlayerCameraTarget
compatibility function
**Function name** EnablePlayerCameraTarget(playerid, enable) **Describe the function** Toggle camera targeting functions for a player. Disabled by default to save bandwidth. **Equivalent in MTA** https://wiki.mtasa.com/wiki/SetCameraTarget **Additional information** https://wiki.sa-mp.com/wiki/EnablePlayerCameraTarget
True
EnablePlayerCameraTarget - **Function name** EnablePlayerCameraTarget(playerid, enable) **Describe the function** Toggle camera targeting functions for a player. Disabled by default to save bandwidth. **Equivalent in MTA** https://wiki.mtasa.com/wiki/SetCameraTarget **Additional information** https://wiki.sa-mp.com/wiki/EnablePlayerCameraTarget
non_process
enableplayercameratarget function name enableplayercameratarget playerid enable describe the function toggle camera targeting functions for a player disabled by default to save bandwidth equivalent in mta additional information
0
14,576
17,702,944,808
IssuesEvent
2021-08-25 01:56:41
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - dcterms:references
Term - change Class - Record-level non-normative Process - complete
## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): Clarity * Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/terms/#dcterms:references, https://dublincore.org/specifications/dublin-core/dcmi-terms/#references Proposed new attributes of the term: * Term name (in lowerCamelCase): dcterms:references * Organized in Class (e.g. Location, Taxon): Record-level * Definition of the term: (unchanged): A related resource that is referenced, cited, or otherwise pointed to by the described resource. * Usage comments (recommendations regarding content, etc.): **From Dublin Core, "This property is intended to be used with non-literal values. This property is an inverse property of Is Referenced By." The intended usage of this term in Darwin Core is to point to the definitive source representation of the resource (e.g.,Taxon, Occurrence, Event in Darwin Core), if one is available. Note that the intended usage of dcterms:bibliographicCitation in Darwin Core, by contrast, is to provide the preferred way to cite the resource itself.** * Examples: **MaterialSample example:**`http://arctos.database.museum/guid/MVZ:Mamm:165861`, **Taxon example: `https://www.catalogueoflife.org/data/taxon/32664`** * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): Not in ABCD Note that the definition of this term is governed by the Dublin Core Metadata Initiative.
1.0
Change term - dcterms:references - ## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): Clarity * Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/terms/#dcterms:references, https://dublincore.org/specifications/dublin-core/dcmi-terms/#references Proposed new attributes of the term: * Term name (in lowerCamelCase): dcterms:references * Organized in Class (e.g. Location, Taxon): Record-level * Definition of the term: (unchanged): A related resource that is referenced, cited, or otherwise pointed to by the described resource. * Usage comments (recommendations regarding content, etc.): **From Dublin Core, "This property is intended to be used with non-literal values. This property is an inverse property of Is Referenced By." The intended usage of this term in Darwin Core is to point to the definitive source representation of the resource (e.g.,Taxon, Occurrence, Event in Darwin Core), if one is available. Note that the intended usage of dcterms:bibliographicCitation in Darwin Core, by contrast, is to provide the preferred way to cite the resource itself.** * Examples: **MaterialSample example:**`http://arctos.database.museum/guid/MVZ:Mamm:165861`, **Taxon example: `https://www.catalogueoflife.org/data/taxon/32664`** * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): Not in ABCD Note that the definition of this term is governed by the Dublin Core Metadata Initiative.
process
change term dcterms references change term submitter john wieczorek justification why is this change necessary clarity proponents who needs this change everyone current term definition proposed new attributes of the term term name in lowercamelcase dcterms references organized in class e g location taxon record level definition of the term unchanged a related resource that is referenced cited or otherwise pointed to by the described resource usage comments recommendations regarding content etc from dublin core this property is intended to be used with non literal values this property is an inverse property of is referenced by the intended usage of this term in darwin core is to point to the definitive source representation of the resource e g taxon occurrence event in darwin core if one is available note that the intended usage of dcterms bibliographiccitation in darwin core by contrast is to provide the preferred way to cite the resource itself examples materialsample example taxon example refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable none abcd xpath of the equivalent term in abcd or efg if applicable not in abcd note that the definition of this term is governed by the dublin core metadata initiative
1
12,894
15,283,995,212
IssuesEvent
2021-02-23 11:36:50
2i2c-org/pilot-hubs
https://api.github.com/repos/2i2c-org/pilot-hubs
closed
Create a plan to track hub onboarding, running hubs, and to-do items
team-process
# Background We need a pipeline / workflow around deploying, customizing, and operating new hubs. This should ensure that: - Hubs that need action are addressed quickly - That the hub infrastructure is stable across all hubs (e.g., downtime is minimized) - That any support promises we made can be fulfilled (e.g., responses to issues etc) - That this work is distributed in an equitable fashion amongst 2i2c team members - That this is sustainable for us # ToDo - [ ] Define the process from "this person wants a hub" to "we now have a hub running". (see https://github.com/2i2c-org/pilot-hubs/issues/218 for one pattern we could use) - [ ] Define where we keep track of running hubs, important metadata about them, and conversation about them amongst ourselves and between us and the users (@yuvipanda mentioned that GitHub issues may not be optimal, so we should explore other ideas!) - [ ] Define a process for how 2i2c engineers keep tabs on "to do" items for the hubs, and how they divide labor in tackling them
1.0
Create a plan to track hub onboarding, running hubs, and to-do items - # Background We need a pipeline / workflow around deploying, customizing, and operating new hubs. This should ensure that: - Hubs that need action are addressed quickly - That the hub infrastructure is stable across all hubs (e.g., downtime is minimized) - That any support promises we made can be fulfilled (e.g., responses to issues etc) - That this work is distributed in an equitable fashion amongst 2i2c team members - That this is sustainable for us # ToDo - [ ] Define the process from "this person wants a hub" to "we now have a hub running". (see https://github.com/2i2c-org/pilot-hubs/issues/218 for one pattern we could use) - [ ] Define where we keep track of running hubs, important metadata about them, and conversation about them amongst ourselves and between us and the users (@yuvipanda mentioned that GitHub issues may not be optimal, so we should explore other ideas!) - [ ] Define a process for how 2i2c engineers keep tabs on "to do" items for the hubs, and how they divide labor in tackling them
process
create a plan to track hub onboarding running hubs and to do items background we need a pipeline workflow around deploying customizing and operating new hubs this should ensure that hubs that need action are addressed quickly that the hub infrastructure is stable across all hubs e g downtime is minimized that any support promises we made can be fulfilled e g responses to issues etc that this work is distributed in an equitable fashion amongst team members that this is sustainable for us todo define the process from this person wants a hub to we now have a hub running see for one pattern we could use define where we keep track of running hubs important metadata about them and conversation about them amongst ourselves and between us and the users yuvipanda mentioned that github issues may not be optimal so we should explore other ideas define a process for how engineers keep tabs on to do items for the hubs and how they divide labor in tackling them
1
19,487
25,798,267,635
IssuesEvent
2022-12-10 19:32:11
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
closed
Order of titlepage and abstract
enhancement postprocessing schema
I'm trying to use the `titlepage` environment so that I can use `\title` to define a short title for LaTeXML to use in the HTML title and navigation links. It appears strange to me that the `titlepage` content is generated after the `abstract`: ```tex \documentclass{report} \title{Title provided for LaTeXML} \begin{document} \begin{titlepage} \huge This is the titlepage. \end{titlepage} \begin{abstract} Abstract goes here. \end{abstract} This follows the abstract. \end{document} ``` I would have expected something more like the output when using `\maketitle`: ```tex \documentclass{report} \title{This is the titlepage.} \begin{document} \maketitle \begin{abstract} Abstract goes here. \end{abstract} This follows the abstract. \end{document} ```
1.0
Order of titlepage and abstract - I'm trying to use the `titlepage` environment so that I can use `\title` to define a short title for LaTeXML to use in the HTML title and navigation links. It appears strange to me that the `titlepage` content is generated after the `abstract`: ```tex \documentclass{report} \title{Title provided for LaTeXML} \begin{document} \begin{titlepage} \huge This is the titlepage. \end{titlepage} \begin{abstract} Abstract goes here. \end{abstract} This follows the abstract. \end{document} ``` I would have expected something more like the output when using `\maketitle`: ```tex \documentclass{report} \title{This is the titlepage.} \begin{document} \maketitle \begin{abstract} Abstract goes here. \end{abstract} This follows the abstract. \end{document} ```
process
order of titlepage and abstract i m trying to use the titlepage environment so that i can use title to define a short title for latexml to use in the html title and navigation links it appears strange to me that the titlepage content is generated after the abstract tex documentclass report title title provided for latexml begin document begin titlepage huge this is the titlepage end titlepage begin abstract abstract goes here end abstract this follows the abstract end document i would have expected something more like the output when using maketitle tex documentclass report title this is the titlepage begin document maketitle begin abstract abstract goes here end abstract this follows the abstract end document
1
3,111
6,134,881,646
IssuesEvent
2017-06-26 02:18:48
gaocegege/Processing.R
https://api.github.com/repos/gaocegege/Processing.R
opened
docs: Reference Documentation
community/processing for-new-contributors priority/p3 size/large status/to-be-claimed type/enhancement
- [x] Shape - [x] Lights, Camera - [x] Transform - [ ] Structure - [ ] Environment - [ ] Data - [ ] Control - [ ] Input - [ ] Output - [ ] Image - [ ] Rendering - [ ] Typography - [ ] Math - [ ] Constants
1.0
docs: Reference Documentation - - [x] Shape - [x] Lights, Camera - [x] Transform - [ ] Structure - [ ] Environment - [ ] Data - [ ] Control - [ ] Input - [ ] Output - [ ] Image - [ ] Rendering - [ ] Typography - [ ] Math - [ ] Constants
process
docs reference documentation shape lights camera transform structure environment data control input output image rendering typography math constants
1
19,479
25,789,869,559
IssuesEvent
2022-12-10 02:00:54
googleapis/gapic-generator-java
https://api.github.com/repos/googleapis/gapic-generator-java
opened
Warning: a recent release failed
type: process
The following release PRs may have failed: * #1109 - The release job failed -- check the build log. * #1067 - The release job was triggered, but has not reported back success. * #1079 - The release job was triggered, but has not reported back success. * #1081 - The release job was triggered, but has not reported back success.
1.0
Warning: a recent release failed - The following release PRs may have failed: * #1109 - The release job failed -- check the build log. * #1067 - The release job was triggered, but has not reported back success. * #1079 - The release job was triggered, but has not reported back success. * #1081 - The release job was triggered, but has not reported back success.
process
warning a recent release failed the following release prs may have failed the release job failed check the build log the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success
1
121,999
10,209,709,323
IssuesEvent
2019-08-14 13:20:28
MyBug18/To-The-Infinity
https://api.github.com/repos/MyBug18/To-The-Infinity
opened
TODO list
Test coremodel
- [ ] Add a constructor to Planet class. - [ ] Write a test code. - [ ] Make a simple UI and test job movement.
1.0
TODO list - - [ ] Add a constructor to Planet class. - [ ] Write a test code. - [ ] Make a simple UI and test job movement.
non_process
todo list add a constructor to planet class write a test code make a simple ui and test job movement
0
51,601
13,207,533,716
IssuesEvent
2020-08-14 23:29:04
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
libX11/Xft/Xext not found by ROOT port configure script on some (newer) linux distros (Trac #682)
Incomplete Migration Migrated from Trac defect tools/ports
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/682">https://code.icecube.wisc.edu/projects/icecube/ticket/682</a>, reported by boersmaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T19:07:59", "_ts": "1351710479000000", "description": "On linux Mint 12 the ROOT install script is not able to find the libX11.so, libXft.so and libXext.so libraries, because only a limited list of directories is tested. On my system, libX11.so (and Xft, Xext) live under /usr/lib/i386-linux-gnu, which is not included in that list. I thought of just adding it to the list, which would solve my problem, but I suffered from the dangerous illusion that I could also solve it in an intelligent way, namely by querying the libdir using \"pkg-config\". I tested on 3 different linux systems that pkg-config was installed and gives the correct answer.\n\nAfter applying the following patch to the main configure script, ROOT got built on my system.\n\n{{{\n--- old/configure\t2011-11-24 17:11:19.000000000 +0100\n+++ new/configure\t2012-06-01 16:05:06.001012273 +0200\n@@ -2155,8 +2155,14 @@\n # Check for X11 and Xpm library (for X11 use always shared lib)\n #\n if test ! \"x$enable_x11\" = \"xno\" ; then\n+ if type pkg-config; then\n+\tpkgcnf_libx11dir=`pkg-config --variable=libdir x11`\n+\tpkgcnf_libxftdir=`pkg-config --variable=libdir xft`\n+\tpkgcnf_libxextdir=`pkg-config --variable=libdir xext`\n+ fi\n check_library \"libX11\" \"yes\" \"$x11libdir\" \\\n ${finkdir:+$finkdir/lib} \\\n+\t$pkgcnf_libx11dir \\\n /usr/lib /usr/X11R6/lib /usr/lib/X11 /usr/openwin/lib \\\n /usr/local/lib /usr/local/lib/X11 /usr/local/X11R6/lib \\\n /usr/X11/lib /usr/lib/X11R5\n@@ -2234,6 +2240,7 @@\n \n check_library \"libXft\" \"$enable_shared\" \"$xftlibdir\" \\\n ${finkdir:+$finkdir/lib} \\\n+\t$pkgcnf_libxftdir \\\n /usr/local/lib /usr/lib /usr/lib/X11 \\\n /usr/local/lib/X11 /usr/X11R6/lib /usr/local/X11R6/lib \\\n /usr/X11/lib\n@@ -2246,6 +2253,7 @@\n \n check_library \"libXext\" \"$enable_shared\" \"$xextlibdir\" \\\n ${finkdir:+$finkdir/lib} \\\n+\t$pkgcnf_libxextdir \\\n /usr/local/lib /usr/lib /usr/lib/X11 \\\n /usr/local/lib/X11 /usr/X11R6/lib /usr/local/X11R6/lib \\\n /usr/X11/lib\n}}}", "reporter": "boersma", "cc": "", "resolution": "fixed", "time": "2012-06-01T14:27:28", "component": "tools/ports", "summary": "libX11/Xft/Xext not found by ROOT port configure script on some (newer) linux distros", "priority": "normal", "keywords": "libX11 configure darwinports port ROOT", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
libX11/Xft/Xext not found by ROOT port configure script on some (newer) linux distros (Trac #682) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/682">https://code.icecube.wisc.edu/projects/icecube/ticket/682</a>, reported by boersmaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T19:07:59", "_ts": "1351710479000000", "description": "On linux Mint 12 the ROOT install script is not able to find the libX11.so, libXft.so and libXext.so libraries, because only a limited list of directories is tested. On my system, libX11.so (and Xft, Xext) live under /usr/lib/i386-linux-gnu, which is not included in that list. I thought of just adding it to the list, which would solve my problem, but I suffered from the dangerous illusion that I could also solve it in an intelligent way, namely by querying the libdir using \"pkg-config\". I tested on 3 different linux systems that pkg-config was installed and gives the correct answer.\n\nAfter applying the following patch to the main configure script, ROOT got built on my system.\n\n{{{\n--- old/configure\t2011-11-24 17:11:19.000000000 +0100\n+++ new/configure\t2012-06-01 16:05:06.001012273 +0200\n@@ -2155,8 +2155,14 @@\n # Check for X11 and Xpm library (for X11 use always shared lib)\n #\n if test ! \"x$enable_x11\" = \"xno\" ; then\n+ if type pkg-config; then\n+\tpkgcnf_libx11dir=`pkg-config --variable=libdir x11`\n+\tpkgcnf_libxftdir=`pkg-config --variable=libdir xft`\n+\tpkgcnf_libxextdir=`pkg-config --variable=libdir xext`\n+ fi\n check_library \"libX11\" \"yes\" \"$x11libdir\" \\\n ${finkdir:+$finkdir/lib} \\\n+\t$pkgcnf_libx11dir \\\n /usr/lib /usr/X11R6/lib /usr/lib/X11 /usr/openwin/lib \\\n /usr/local/lib /usr/local/lib/X11 /usr/local/X11R6/lib \\\n /usr/X11/lib /usr/lib/X11R5\n@@ -2234,6 +2240,7 @@\n \n check_library \"libXft\" \"$enable_shared\" \"$xftlibdir\" \\\n ${finkdir:+$finkdir/lib} \\\n+\t$pkgcnf_libxftdir \\\n /usr/local/lib /usr/lib /usr/lib/X11 \\\n /usr/local/lib/X11 /usr/X11R6/lib /usr/local/X11R6/lib \\\n /usr/X11/lib\n@@ -2246,6 +2253,7 @@\n \n check_library \"libXext\" \"$enable_shared\" \"$xextlibdir\" \\\n ${finkdir:+$finkdir/lib} \\\n+\t$pkgcnf_libxextdir \\\n /usr/local/lib /usr/lib /usr/lib/X11 \\\n /usr/local/lib/X11 /usr/X11R6/lib /usr/local/X11R6/lib \\\n /usr/X11/lib\n}}}", "reporter": "boersma", "cc": "", "resolution": "fixed", "time": "2012-06-01T14:27:28", "component": "tools/ports", "summary": "libX11/Xft/Xext not found by ROOT port configure script on some (newer) linux distros", "priority": "normal", "keywords": "libX11 configure darwinports port ROOT", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
non_process
xft xext not found by root port configure script on some newer linux distros trac migrated from json status closed changetime ts description on linux mint the root install script is not able to find the so libxft so and libxext so libraries because only a limited list of directories is tested on my system so and xft xext live under usr lib linux gnu which is not included in that list i thought of just adding it to the list which would solve my problem but i suffered from the dangerous illusion that i could also solve it in an intelligent way namely by querying the libdir using pkg config i tested on different linux systems that pkg config was installed and gives the correct answer n nafter applying the following patch to the main configure script root got built on my system n n n old configure n new configure n n check for and xpm library for use always shared lib n n if test x enable xno then n if type pkg config then n tpkgcnf pkg config variable libdir n tpkgcnf libxftdir pkg config variable libdir xft n tpkgcnf libxextdir pkg config variable libdir xext n fi n check library yes n finkdir finkdir lib n t pkgcnf n usr lib usr lib usr lib usr openwin lib n usr local lib usr local lib usr local lib n usr lib usr lib n n n check library libxft enable shared xftlibdir n finkdir finkdir lib n t pkgcnf libxftdir n usr local lib usr lib usr lib n usr local lib usr lib usr local lib n usr lib n n n check library libxext enable shared xextlibdir n finkdir finkdir lib n t pkgcnf libxextdir n usr local lib usr lib usr lib n usr local lib usr lib usr local lib n usr lib n reporter boersma cc resolution fixed time component tools ports summary xft xext not found by root port configure script on some newer linux distros priority normal keywords configure darwinports port root milestone owner nega type defect
0
13,005
15,362,429,338
IssuesEvent
2021-03-01 19:27:56
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
chore: use multiple projects for datastore/cloud-client
priority: p2 samples type: process
I saw the following error: ``` google.api_core.exceptions.Aborted: 409 too much contention on these datastore entities. please try again. entity groups: [(app=s~python-docs-samples-tests, Task, 5189114174373888)] ``` in one presubmit build. Maybe we should use multiple projects. It's rather involving as follows: - need to initialize Datastore on the project - need to create bunch of composite indices - need to set appropriate permission on the test service accounts
1.0
chore: use multiple projects for datastore/cloud-client - I saw the following error: ``` google.api_core.exceptions.Aborted: 409 too much contention on these datastore entities. please try again. entity groups: [(app=s~python-docs-samples-tests, Task, 5189114174373888)] ``` in one presubmit build. Maybe we should use multiple projects. It's rather involving as follows: - need to initialize Datastore on the project - need to create bunch of composite indices - need to set appropriate permission on the test service accounts
process
chore use multiple projects for datastore cloud client i saw the following error google api core exceptions aborted too much contention on these datastore entities please try again entity groups in one presubmit build maybe we should use multiple projects it s rather involving as follows need to initialize datastore on the project need to create bunch of composite indices need to set appropriate permission on the test service accounts
1
97,965
29,146,431,785
IssuesEvent
2023-05-18 03:40:41
facebookincubator/velox
https://api.github.com/repos/facebookincubator/velox
closed
linux-build circle CI build failing
build broken-build
### Problem description looks like we are hitting memory limit while linking velox in our circleCI builds https://app.circleci.com/pipelines/github/facebookincubator/velox/22844/workflows/b15ce282-0ba3-4565-ba17-9cc166ee9bd5/jobs/144524 Looking at the resource usage of the instance confirms RAM usage going upto 99% and then drops rapidly (cause the linked failed I presume). Searching online and solutions are: 1. either increase the RAM (not sure if we can do that, as we would have to change the instance type) 2. Restrict the build to use less memory, like building in release mode 3. increasing swap space (again not sure how to do this) Any suggestions would be appreciated ### System information circleCI linux-build job: https://app.circleci.com/pipelines/github/facebookincubator/velox/22844/workflows/b15ce282-0ba3-4565-ba17-9cc166ee9bd5/jobs/144524 ### CMake log ```bash FAILED: velox/functions/prestosql/aggregates/tests/velox_aggregates_test : && /opt/rh/gcc-toolset-9/root/bin/g++ -mavx2 -mfma -mavx -mf16c -mlzcnt -std=c++17 -mbmi2 -D USE_VELOX_COMMON_BASE -D HAS_UNCAUGHT_EXCEPTIONS -Wall -Wextra -Wno-unused -Wno-unused-parameter -Wno-sign-compare -Wno-ignored-qualifiers -Wno-implicit-fallthrough -Wno-empty-body -Wno-class-memaccess -Wno-comment -Wno-int-in-bool-context -Wno-redundant-move -Wno-array-bounds -Wno-maybe-uninitialized -Wno-unused-result -Wno-format-overflow -Wno-strict-aliasing -Wno-type-limits -Wno-stringop-overflow -Werror -g -Wl,-export-dynamic velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HiveConfig.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HiveConnector.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HiveDataSink.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HivePartitionUtil.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/FileHandle.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/PartitionIdGenerator.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/AggregationFunctionRegTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ApproxDistinctTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ApproxMostFrequentTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ApproxPercentileTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ArbitraryTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ArrayAggTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/AverageAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/BitwiseAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/BoolAndOrTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ChecksumAggregateTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/CountAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/CountIfAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/CovarianceAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/HistogramTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/Main.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MinMaxByAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MinMaxTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/PrestoHasherTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/SumTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MapAggTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MapUnionAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MapUnionSumTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ValueListTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/VarianceAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MaxSizeForStatsTest.cpp.o -o velox/functions/prestosql/aggregates/tests/velox_aggregates_test -L/root/project/_build/debug/_deps/icu/lib -Wl,-rpath,/root/project/_build/debug/_deps/icu/lib:/root/project/_build/debug/_deps/boost-build/libs/context:/root/project/_build/debug/_deps/boost-build/libs/filesystem:/root/project/_build/debug/_deps/boost-build/libs/program_options:/root/project/_build/debug/_deps/boost-build/libs/thread:/root/project/_build/debug/_deps/boost-build/libs/atomic:/root/project/_build/debug/_deps/boost-build/libs/date_time:/root/project/_build/debug/_deps/boost-build/libs/chrono:/root/project/_build/debug/_deps/boost-build/libs/container velox/functions/prestosql/aggregates/libvelox_aggregates.a velox/functions/prestosql/aggregates/tests/libvelox_aggregates_test_lib.a velox/core/libvelox_core.a velox/dwio/common/tests/utils/libvelox_dwio_common_test_utils.a velox/exec/libvelox_exec.a velox/exec/tests/utils/libvelox_exec_test_lib.a velox/common/file/libvelox_file.a velox/functions/prestosql/tests/utils/libvelox_functions_test_lib.a velox/functions/prestosql/registration/libvelox_functions_prestosql.a velox/functions/lib/libvelox_functions_lib.a velox/type/libvelox_type.a velox/vector/fuzzer/libvelox_vector_fuzzer.a _deps/gflags-build/libgflags_debug.a lib/libgtest.a lib/libgtest_main.a velox/exec/tests/utils/libvelox_exec_test_lib.a velox/functions/prestosql/aggregates/libvelox_aggregates.a velox/dwio/common/tests/utils/libvelox_dwio_common_test_utils.a velox/functions/prestosql/registration/libvelox_functions_prestosql.a velox/functions/prestosql/libvelox_functions_prestosql_impl.a velox/functions/lib/libvelox_functions_lib.a _deps/re2-build/libre2.a velox/common/hyperloglog/libvelox_common_hyperloglog.a velox/functions/prestosql/json/libvelox_functions_json.a velox/external/md5/libmd5.a velox/functions/prestosql/types/libvelox_presto_types.a velox/functions/lib/libvelox_is_null_functions.a velox/serializers/libvelox_presto_serializer.a velox/dwio/type/fbhive/libvelox_dwio_type_fbhive.a velox/exec/tests/utils/libvelox_temp_path.a velox/dwio/dwrf/reader/libvelox_dwio_dwrf_reader.a velox/dwio/dwrf/writer/libvelox_dwio_dwrf_writer.a velox/dwio/dwrf/common/libvelox_dwio_dwrf_common.a velox/dwio/common/libvelox_dwio_common.a velox/dwio/common/exception/libvelox_dwio_common_exception.a velox/dwio/common/encryption/libvelox_dwio_common_encryption.a velox/common/caching/libvelox_caching.a velox/dwio/common/compression/libvelox_dwio_common_compression.a velox/dwio/dwrf/utils/libvelox_dwio_dwrf_utils.a velox/dwio/dwrf/proto/libvelox_dwio_dwrf_proto.a _deps/protobuf-build/libprotobufd.a /usr/lib64/liblzo2.so velox/tpch/gen/libvelox_tpch_gen.a velox/vector/tests/utils/libvelox_vector_test_lib.a velox/exec/libvelox_exec.a velox/codegen/libvelox_codegen.a velox/exec/libvelox_exec.a velox/codegen/libvelox_codegen.a velox/common/file/libvelox_file.a velox/connectors/libvelox_connector.a velox/vector/arrow/libvelox_arrow_bridge.a velox/parse/libvelox_parse_parser.a velox/duckdb/conversion/libvelox_duckdb_parser.a velox/duckdb/conversion/libvelox_duckdb_conversion.a velox/external/duckdb/tpch/libtpch_extension.a velox/external/duckdb/tpch/dbgen/libdbgen.a velox/parse/libvelox_parse_expression.a velox/parse/libvelox_parse_utils.a velox/functions/libvelox_function_registry.a velox/expression/libvelox_expression.a velox/core/libvelox_core.a velox/core/libvelox_config.a velox/functions/lib/libvelox_functions_util.a velox/expression/libvelox_expression_functions.a velox/expression/type_calculation/libvelox_type_calculation.a velox/external/duckdb/libduckdb.a velox/vector/libvelox_vector.a velox/type/libvelox_type.a velox/common/serialization/libvelox_serialization.a velox/type/tz/libvelox_type_tz.a velox/external/date/libvelox_external_date.a velox/common/time/libvelox_time.a velox/common/encode/libvelox_encode.a velox/buffer/libvelox_buffer.a velox/common/memory/libvelox_memory.a velox/common/testutil/libvelox_test_util.a velox/common/base/libvelox_common_base.a velox/common/base/libvelox_exception.a velox/common/process/libvelox_process.a _deps/folly-build/libfolly.a _deps/fmt-build/libfmtd.a _deps/glog-build/libglogd.a /usr/lib64/libz.so /usr/lib64/libssl.so /usr/lib64/libcrypto.so /usr/lib64/liblz4.so /usr/lib64/libzstd.so /usr/lib64/libdwarf.so _deps/boost-build/libs/context/libboost_context.so.1.81.0 _deps/boost-build/libs/filesystem/libboost_filesystem.so.1.81.0 _deps/boost-build/libs/program_options/libboost_program_options.so.1.81.0 _deps/boost-build/libs/thread/libboost_thread.so.1.81.0 _deps/boost-build/libs/atomic/libboost_atomic.so.1.81.0 _deps/boost-build/libs/date_time/libboost_date_time.so.1.81.0 _deps/boost-build/libs/chrono/libboost_chrono.so.1.81.0 _deps/boost-build/libs/container/libboost_container.so.1.81.0 _deps/boost-build/libs/exception/libboost_exception.a /usr/lib64/libdouble-conversion.so /usr/lib64/libevent.so /usr/local/lib64/libsnappy.a -ldl _deps/glog-build/libglogd.a _deps/gflags-build/libgflags_debug.a lib/libgtest.a -pthread && : collect2: fatal error: ld terminated with signal 9 [Killed] compilation terminated. ```
2.0
linux-build circle CI build failing - ### Problem description looks like we are hitting memory limit while linking velox in our circleCI builds https://app.circleci.com/pipelines/github/facebookincubator/velox/22844/workflows/b15ce282-0ba3-4565-ba17-9cc166ee9bd5/jobs/144524 Looking at the resource usage of the instance confirms RAM usage going upto 99% and then drops rapidly (cause the linked failed I presume). Searching online and solutions are: 1. either increase the RAM (not sure if we can do that, as we would have to change the instance type) 2. Restrict the build to use less memory, like building in release mode 3. increasing swap space (again not sure how to do this) Any suggestions would be appreciated ### System information circleCI linux-build job: https://app.circleci.com/pipelines/github/facebookincubator/velox/22844/workflows/b15ce282-0ba3-4565-ba17-9cc166ee9bd5/jobs/144524 ### CMake log ```bash FAILED: velox/functions/prestosql/aggregates/tests/velox_aggregates_test : && /opt/rh/gcc-toolset-9/root/bin/g++ -mavx2 -mfma -mavx -mf16c -mlzcnt -std=c++17 -mbmi2 -D USE_VELOX_COMMON_BASE -D HAS_UNCAUGHT_EXCEPTIONS -Wall -Wextra -Wno-unused -Wno-unused-parameter -Wno-sign-compare -Wno-ignored-qualifiers -Wno-implicit-fallthrough -Wno-empty-body -Wno-class-memaccess -Wno-comment -Wno-int-in-bool-context -Wno-redundant-move -Wno-array-bounds -Wno-maybe-uninitialized -Wno-unused-result -Wno-format-overflow -Wno-strict-aliasing -Wno-type-limits -Wno-stringop-overflow -Werror -g -Wl,-export-dynamic velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HiveConfig.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HiveConnector.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HiveDataSink.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/HivePartitionUtil.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/FileHandle.cpp.o velox/connectors/hive/CMakeFiles/velox_hive_connector.dir/PartitionIdGenerator.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/AggregationFunctionRegTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ApproxDistinctTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ApproxMostFrequentTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ApproxPercentileTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ArbitraryTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ArrayAggTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/AverageAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/BitwiseAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/BoolAndOrTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ChecksumAggregateTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/CountAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/CountIfAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/CovarianceAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/HistogramTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/Main.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MinMaxByAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MinMaxTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/PrestoHasherTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/SumTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MapAggTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MapUnionAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MapUnionSumTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/ValueListTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/VarianceAggregationTest.cpp.o velox/functions/prestosql/aggregates/tests/CMakeFiles/velox_aggregates_test.dir/MaxSizeForStatsTest.cpp.o -o velox/functions/prestosql/aggregates/tests/velox_aggregates_test -L/root/project/_build/debug/_deps/icu/lib -Wl,-rpath,/root/project/_build/debug/_deps/icu/lib:/root/project/_build/debug/_deps/boost-build/libs/context:/root/project/_build/debug/_deps/boost-build/libs/filesystem:/root/project/_build/debug/_deps/boost-build/libs/program_options:/root/project/_build/debug/_deps/boost-build/libs/thread:/root/project/_build/debug/_deps/boost-build/libs/atomic:/root/project/_build/debug/_deps/boost-build/libs/date_time:/root/project/_build/debug/_deps/boost-build/libs/chrono:/root/project/_build/debug/_deps/boost-build/libs/container velox/functions/prestosql/aggregates/libvelox_aggregates.a velox/functions/prestosql/aggregates/tests/libvelox_aggregates_test_lib.a velox/core/libvelox_core.a velox/dwio/common/tests/utils/libvelox_dwio_common_test_utils.a velox/exec/libvelox_exec.a velox/exec/tests/utils/libvelox_exec_test_lib.a velox/common/file/libvelox_file.a velox/functions/prestosql/tests/utils/libvelox_functions_test_lib.a velox/functions/prestosql/registration/libvelox_functions_prestosql.a velox/functions/lib/libvelox_functions_lib.a velox/type/libvelox_type.a velox/vector/fuzzer/libvelox_vector_fuzzer.a _deps/gflags-build/libgflags_debug.a lib/libgtest.a lib/libgtest_main.a velox/exec/tests/utils/libvelox_exec_test_lib.a velox/functions/prestosql/aggregates/libvelox_aggregates.a velox/dwio/common/tests/utils/libvelox_dwio_common_test_utils.a velox/functions/prestosql/registration/libvelox_functions_prestosql.a velox/functions/prestosql/libvelox_functions_prestosql_impl.a velox/functions/lib/libvelox_functions_lib.a _deps/re2-build/libre2.a velox/common/hyperloglog/libvelox_common_hyperloglog.a velox/functions/prestosql/json/libvelox_functions_json.a velox/external/md5/libmd5.a velox/functions/prestosql/types/libvelox_presto_types.a velox/functions/lib/libvelox_is_null_functions.a velox/serializers/libvelox_presto_serializer.a velox/dwio/type/fbhive/libvelox_dwio_type_fbhive.a velox/exec/tests/utils/libvelox_temp_path.a velox/dwio/dwrf/reader/libvelox_dwio_dwrf_reader.a velox/dwio/dwrf/writer/libvelox_dwio_dwrf_writer.a velox/dwio/dwrf/common/libvelox_dwio_dwrf_common.a velox/dwio/common/libvelox_dwio_common.a velox/dwio/common/exception/libvelox_dwio_common_exception.a velox/dwio/common/encryption/libvelox_dwio_common_encryption.a velox/common/caching/libvelox_caching.a velox/dwio/common/compression/libvelox_dwio_common_compression.a velox/dwio/dwrf/utils/libvelox_dwio_dwrf_utils.a velox/dwio/dwrf/proto/libvelox_dwio_dwrf_proto.a _deps/protobuf-build/libprotobufd.a /usr/lib64/liblzo2.so velox/tpch/gen/libvelox_tpch_gen.a velox/vector/tests/utils/libvelox_vector_test_lib.a velox/exec/libvelox_exec.a velox/codegen/libvelox_codegen.a velox/exec/libvelox_exec.a velox/codegen/libvelox_codegen.a velox/common/file/libvelox_file.a velox/connectors/libvelox_connector.a velox/vector/arrow/libvelox_arrow_bridge.a velox/parse/libvelox_parse_parser.a velox/duckdb/conversion/libvelox_duckdb_parser.a velox/duckdb/conversion/libvelox_duckdb_conversion.a velox/external/duckdb/tpch/libtpch_extension.a velox/external/duckdb/tpch/dbgen/libdbgen.a velox/parse/libvelox_parse_expression.a velox/parse/libvelox_parse_utils.a velox/functions/libvelox_function_registry.a velox/expression/libvelox_expression.a velox/core/libvelox_core.a velox/core/libvelox_config.a velox/functions/lib/libvelox_functions_util.a velox/expression/libvelox_expression_functions.a velox/expression/type_calculation/libvelox_type_calculation.a velox/external/duckdb/libduckdb.a velox/vector/libvelox_vector.a velox/type/libvelox_type.a velox/common/serialization/libvelox_serialization.a velox/type/tz/libvelox_type_tz.a velox/external/date/libvelox_external_date.a velox/common/time/libvelox_time.a velox/common/encode/libvelox_encode.a velox/buffer/libvelox_buffer.a velox/common/memory/libvelox_memory.a velox/common/testutil/libvelox_test_util.a velox/common/base/libvelox_common_base.a velox/common/base/libvelox_exception.a velox/common/process/libvelox_process.a _deps/folly-build/libfolly.a _deps/fmt-build/libfmtd.a _deps/glog-build/libglogd.a /usr/lib64/libz.so /usr/lib64/libssl.so /usr/lib64/libcrypto.so /usr/lib64/liblz4.so /usr/lib64/libzstd.so /usr/lib64/libdwarf.so _deps/boost-build/libs/context/libboost_context.so.1.81.0 _deps/boost-build/libs/filesystem/libboost_filesystem.so.1.81.0 _deps/boost-build/libs/program_options/libboost_program_options.so.1.81.0 _deps/boost-build/libs/thread/libboost_thread.so.1.81.0 _deps/boost-build/libs/atomic/libboost_atomic.so.1.81.0 _deps/boost-build/libs/date_time/libboost_date_time.so.1.81.0 _deps/boost-build/libs/chrono/libboost_chrono.so.1.81.0 _deps/boost-build/libs/container/libboost_container.so.1.81.0 _deps/boost-build/libs/exception/libboost_exception.a /usr/lib64/libdouble-conversion.so /usr/lib64/libevent.so /usr/local/lib64/libsnappy.a -ldl _deps/glog-build/libglogd.a _deps/gflags-build/libgflags_debug.a lib/libgtest.a -pthread && : collect2: fatal error: ld terminated with signal 9 [Killed] compilation terminated. ```
non_process
linux build circle ci build failing problem description looks like we are hitting memory limit while linking velox in our circleci builds looking at the resource usage of the instance confirms ram usage going upto and then drops rapidly cause the linked failed i presume searching online and solutions are either increase the ram not sure if we can do that as we would have to change the instance type restrict the build to use less memory like building in release mode increasing swap space again not sure how to do this any suggestions would be appreciated system information circleci linux build job cmake log bash failed velox functions prestosql aggregates tests velox aggregates test opt rh gcc toolset root bin g mfma mavx mlzcnt std c d use velox common base d has uncaught exceptions wall wextra wno unused wno unused parameter wno sign compare wno ignored qualifiers wno implicit fallthrough wno empty body wno class memaccess wno comment wno int in bool context wno redundant move wno array bounds wno maybe uninitialized wno unused result wno format overflow wno strict aliasing wno type limits wno stringop overflow werror g wl export dynamic velox connectors hive cmakefiles velox hive connector dir hiveconfig cpp o velox connectors hive cmakefiles velox hive connector dir hiveconnector cpp o velox connectors hive cmakefiles velox hive connector dir hivedatasink cpp o velox connectors hive cmakefiles velox hive connector dir hivepartitionutil cpp o velox connectors hive cmakefiles velox hive connector dir filehandle cpp o velox connectors hive cmakefiles velox hive connector dir partitionidgenerator cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir aggregationfunctionregtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir approxdistincttest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir approxmostfrequenttest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir approxpercentiletest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir arbitrarytest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir arrayaggtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir averageaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir bitwiseaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir boolandortest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir checksumaggregatetest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir countaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir countifaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir covarianceaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir histogramtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir main cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir minmaxbyaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir minmaxtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir prestohashertest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir sumtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir mapaggtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir mapunionaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir mapunionsumtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir valuelisttest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir varianceaggregationtest cpp o velox functions prestosql aggregates tests cmakefiles velox aggregates test dir maxsizeforstatstest cpp o o velox functions prestosql aggregates tests velox aggregates test l root project build debug deps icu lib wl rpath root project build debug deps icu lib root project build debug deps boost build libs context root project build debug deps boost build libs filesystem root project build debug deps boost build libs program options root project build debug deps boost build libs thread root project build debug deps boost build libs atomic root project build debug deps boost build libs date time root project build debug deps boost build libs chrono root project build debug deps boost build libs container velox functions prestosql aggregates libvelox aggregates a velox functions prestosql aggregates tests libvelox aggregates test lib a velox core libvelox core a velox dwio common tests utils libvelox dwio common test utils a velox exec libvelox exec a velox exec tests utils libvelox exec test lib a velox common file libvelox file a velox functions prestosql tests utils libvelox functions test lib a velox functions prestosql registration libvelox functions prestosql a velox functions lib libvelox functions lib a velox type libvelox type a velox vector fuzzer libvelox vector fuzzer a deps gflags build libgflags debug a lib libgtest a lib libgtest main a velox exec tests utils libvelox exec test lib a velox functions prestosql aggregates libvelox aggregates a velox dwio common tests utils libvelox dwio common test utils a velox functions prestosql registration libvelox functions prestosql a velox functions prestosql libvelox functions prestosql impl a velox functions lib libvelox functions lib a deps build a velox common hyperloglog libvelox common hyperloglog a velox functions prestosql json libvelox functions json a velox external a velox functions prestosql types libvelox presto types a velox functions lib libvelox is null functions a velox serializers libvelox presto serializer a velox dwio type fbhive libvelox dwio type fbhive a velox exec tests utils libvelox temp path a velox dwio dwrf reader libvelox dwio dwrf reader a velox dwio dwrf writer libvelox dwio dwrf writer a velox dwio dwrf common libvelox dwio dwrf common a velox dwio common libvelox dwio common a velox dwio common exception libvelox dwio common exception a velox dwio common encryption libvelox dwio common encryption a velox common caching libvelox caching a velox dwio common compression libvelox dwio common compression a velox dwio dwrf utils libvelox dwio dwrf utils a velox dwio dwrf proto libvelox dwio dwrf proto a deps protobuf build libprotobufd a usr so velox tpch gen libvelox tpch gen a velox vector tests utils libvelox vector test lib a velox exec libvelox exec a velox codegen libvelox codegen a velox exec libvelox exec a velox codegen libvelox codegen a velox common file libvelox file a velox connectors libvelox connector a velox vector arrow libvelox arrow bridge a velox parse libvelox parse parser a velox duckdb conversion libvelox duckdb parser a velox duckdb conversion libvelox duckdb conversion a velox external duckdb tpch libtpch extension a velox external duckdb tpch dbgen libdbgen a velox parse libvelox parse expression a velox parse libvelox parse utils a velox functions libvelox function registry a velox expression libvelox expression a velox core libvelox core a velox core libvelox config a velox functions lib libvelox functions util a velox expression libvelox expression functions a velox expression type calculation libvelox type calculation a velox external duckdb libduckdb a velox vector libvelox vector a velox type libvelox type a velox common serialization libvelox serialization a velox type tz libvelox type tz a velox external date libvelox external date a velox common time libvelox time a velox common encode libvelox encode a velox buffer libvelox buffer a velox common memory libvelox memory a velox common testutil libvelox test util a velox common base libvelox common base a velox common base libvelox exception a velox common process libvelox process a deps folly build libfolly a deps fmt build libfmtd a deps glog build libglogd a usr libz so usr libssl so usr libcrypto so usr so usr libzstd so usr libdwarf so deps boost build libs context libboost context so deps boost build libs filesystem libboost filesystem so deps boost build libs program options libboost program options so deps boost build libs thread libboost thread so deps boost build libs atomic libboost atomic so deps boost build libs date time libboost date time so deps boost build libs chrono libboost chrono so deps boost build libs container libboost container so deps boost build libs exception libboost exception a usr libdouble conversion so usr libevent so usr local libsnappy a ldl deps glog build libglogd a deps gflags build libgflags debug a lib libgtest a pthread fatal error ld terminated with signal compilation terminated
0
554,395
16,419,588,095
IssuesEvent
2021-05-19 10:55:28
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
wwwmat.sat.gob.mx - site is not usable
browser-firefox-ios ml-needsdiagnosis-false os-ios priority-normal
<!-- @browser: Firefox iOS 33.1 --> <!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_5_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/74294 --> **URL**: https://wwwmat.sat.gob.mx/Apps/default.html **Browser / Version**: Firefox iOS 33.1 **Operating System**: iOS 14.5.1 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: I came to this portal, introduced user account and password, then never got a response, several reloads tried, no response. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
wwwmat.sat.gob.mx - site is not usable - <!-- @browser: Firefox iOS 33.1 --> <!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_5_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/74294 --> **URL**: https://wwwmat.sat.gob.mx/Apps/default.html **Browser / Version**: Firefox iOS 33.1 **Operating System**: iOS 14.5.1 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: I came to this portal, introduced user account and password, then never got a response, several reloads tried, no response. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
wwwmat sat gob mx site is not usable url browser version firefox ios operating system ios tested another browser no problem type site is not usable description page not loading correctly steps to reproduce i came to this portal introduced user account and password then never got a response several reloads tried no response browser configuration none from with ❤️
0
109,572
13,786,554,159
IssuesEvent
2020-10-09 02:12:14
codethesaurus/codethesaur.us
https://api.github.com/repos/codethesaurus/codethesaur.us
closed
Add another couple of concepts to thesauruses
Pre-MVP setup thesaurus meta design
Maybe add some files for functions and classes to help give some more dimensions to how the product will work.
1.0
Add another couple of concepts to thesauruses - Maybe add some files for functions and classes to help give some more dimensions to how the product will work.
non_process
add another couple of concepts to thesauruses maybe add some files for functions and classes to help give some more dimensions to how the product will work
0
424,134
12,306,262,447
IssuesEvent
2020-05-12 00:54:58
momentum-mod/game
https://api.github.com/repos/momentum-mod/game
reopened
Add HL2-style AH (Accelerated Hopping) Gamemode
Priority: Low Size: Large Type: Feature
There's been an increased interest in running custom HL2 maps lately, so myself and Waezone have been looking to get the game's splendidly wacky movement in Momentum. We've managed to get the essential movement working almost perfectly in under 200 lines of new code, and the only thing we've had to change in code related to other gamemodes is explicitly set their gravity to 800, as we set it to 650 in AH. - [x] Add the new gamemode, auto set movement vars etc. - **Done** - [x] Implement the core HL2 movement - [x] ABH/ASH/AFH - **Almost done** - Apart from the speed of crouched AHs being slightly off (we think we know why this is off, will update tomorrow) - [x] Sprint - **WIP** - We have HL2-style implementation, but it's clunky and can be improved (namely, all it means to be sprinting is for bIsSprinting to be true on each tick. This is a pain if we want to call stuff when the player starts and stops sprinting.) - [x] Crouch - **Done** Mechanics for this are working fine - [x] Toggleduck - **Todo** The final, really niche bit of movement tech - [ ] Add togglable suit sprint noises - **Todo** - [ ] Start zone speed limit to 700 ups - [x] Add sprint and walk to keypress UI - **Todo** Our fork can be found here https://github.com/tsa96/game/tree/feature/hl2-movement A zoned AH map, made by Mac, can be found here: http://trumpetdust.org/dl/ah_orange.bsp http://trumpetdust.org/dl/ah_orange.zon
1.0
Add HL2-style AH (Accelerated Hopping) Gamemode - There's been an increased interest in running custom HL2 maps lately, so myself and Waezone have been looking to get the game's splendidly wacky movement in Momentum. We've managed to get the essential movement working almost perfectly in under 200 lines of new code, and the only thing we've had to change in code related to other gamemodes is explicitly set their gravity to 800, as we set it to 650 in AH. - [x] Add the new gamemode, auto set movement vars etc. - **Done** - [x] Implement the core HL2 movement - [x] ABH/ASH/AFH - **Almost done** - Apart from the speed of crouched AHs being slightly off (we think we know why this is off, will update tomorrow) - [x] Sprint - **WIP** - We have HL2-style implementation, but it's clunky and can be improved (namely, all it means to be sprinting is for bIsSprinting to be true on each tick. This is a pain if we want to call stuff when the player starts and stops sprinting.) - [x] Crouch - **Done** Mechanics for this are working fine - [x] Toggleduck - **Todo** The final, really niche bit of movement tech - [ ] Add togglable suit sprint noises - **Todo** - [ ] Start zone speed limit to 700 ups - [x] Add sprint and walk to keypress UI - **Todo** Our fork can be found here https://github.com/tsa96/game/tree/feature/hl2-movement A zoned AH map, made by Mac, can be found here: http://trumpetdust.org/dl/ah_orange.bsp http://trumpetdust.org/dl/ah_orange.zon
non_process
add style ah accelerated hopping gamemode there s been an increased interest in running custom maps lately so myself and waezone have been looking to get the game s splendidly wacky movement in momentum we ve managed to get the essential movement working almost perfectly in under lines of new code and the only thing we ve had to change in code related to other gamemodes is explicitly set their gravity to as we set it to in ah add the new gamemode auto set movement vars etc done implement the core movement abh ash afh almost done apart from the speed of crouched ahs being slightly off we think we know why this is off will update tomorrow sprint wip we have style implementation but it s clunky and can be improved namely all it means to be sprinting is for bissprinting to be true on each tick this is a pain if we want to call stuff when the player starts and stops sprinting crouch done mechanics for this are working fine toggleduck todo the final really niche bit of movement tech add togglable suit sprint noises todo start zone speed limit to ups add sprint and walk to keypress ui todo our fork can be found here a zoned ah map made by mac can be found here
0
204,062
23,203,421,992
IssuesEvent
2022-08-02 01:06:54
swagger-api/swagger-codegen
https://api.github.com/repos/swagger-api/swagger-codegen
reopened
CVE-2021-44906 (High) detected in multiple libraries
security vulnerability
## CVE-2021-44906 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-1.2.0.tgz</b>, <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.5.tgz</b></p></summary> <p> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p> Dependency Hierarchy: - ts-node-3.3.0.tgz (Root Library) - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /samples/client/petstore/javascript-promise-es6/package.json</p> <p>Path to vulnerable library: /samples/client/petstore/javascript-promise-es6/node_modules/minimist/package.json,/samples/dynamic-html/node_modules/minimist/package.json,/samples/client/petstore/typescript-fetch/tests/default/node_modules/mkdirp/node_modules/minimist/package.json,/samples/client/petstore/javascript-es6/node_modules/minimist/package.json,/samples/client/petstore/typescript-node/npm/node_modules/mkdirp/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - express-3.21.2.tgz (Root Library) - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.5.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p> <p>Path to dependency file: /samples/server/petstore/nodejs/package.json</p> <p>Path to vulnerable library: /samples/server/petstore/nodejs/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - swagger-tools-0.10.1.tgz (Root Library) - multer-1.4.2.tgz - mkdirp-0.5.5.tgz - :x: **minimist-1.2.5.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/swagger-api/swagger-codegen/commit/4b7a8d7d7384aa6a27d6309c35ade0916edae7ed">4b7a8d7d7384aa6a27d6309c35ade0916edae7ed</a></p> <p>Found in base branches: <b>3.0.0, master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95). <p>Publish Date: 2022-03-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-03-17</p> <p>Fix Resolution (minimist): 1.2.6</p> <p>Direct dependency fix Resolution (ts-node): 8.0.0</p><p>Fix Resolution (minimist): 1.2.6</p> <p>Direct dependency fix Resolution (express): 4.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2021-44906 (High) detected in multiple libraries - ## CVE-2021-44906 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-1.2.0.tgz</b>, <b>minimist-0.0.8.tgz</b>, <b>minimist-1.2.5.tgz</b></p></summary> <p> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p> Dependency Hierarchy: - ts-node-3.3.0.tgz (Root Library) - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p>Path to dependency file: /samples/client/petstore/javascript-promise-es6/package.json</p> <p>Path to vulnerable library: /samples/client/petstore/javascript-promise-es6/node_modules/minimist/package.json,/samples/dynamic-html/node_modules/minimist/package.json,/samples/client/petstore/typescript-fetch/tests/default/node_modules/mkdirp/node_modules/minimist/package.json,/samples/client/petstore/javascript-es6/node_modules/minimist/package.json,/samples/client/petstore/typescript-node/npm/node_modules/mkdirp/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - express-3.21.2.tgz (Root Library) - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-1.2.5.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p> <p>Path to dependency file: /samples/server/petstore/nodejs/package.json</p> <p>Path to vulnerable library: /samples/server/petstore/nodejs/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - swagger-tools-0.10.1.tgz (Root Library) - multer-1.4.2.tgz - mkdirp-0.5.5.tgz - :x: **minimist-1.2.5.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/swagger-api/swagger-codegen/commit/4b7a8d7d7384aa6a27d6309c35ade0916edae7ed">4b7a8d7d7384aa6a27d6309c35ade0916edae7ed</a></p> <p>Found in base branches: <b>3.0.0, master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95). <p>Publish Date: 2022-03-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44906>CVE-2021-44906</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-03-17</p> <p>Fix Resolution (minimist): 1.2.6</p> <p>Direct dependency fix Resolution (ts-node): 8.0.0</p><p>Fix Resolution (minimist): 1.2.6</p> <p>Direct dependency fix Resolution (express): 4.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_process
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz minimist tgz parse argument options library home page a href dependency hierarchy ts node tgz root library x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file samples client petstore javascript promise package json path to vulnerable library samples client petstore javascript promise node modules minimist package json samples dynamic html node modules minimist package json samples client petstore typescript fetch tests default node modules mkdirp node modules minimist package json samples client petstore javascript node modules minimist package json samples client petstore typescript node npm node modules mkdirp node modules minimist package json dependency hierarchy express tgz root library mkdirp tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href path to dependency file samples server petstore nodejs package json path to vulnerable library samples server petstore nodejs node modules minimist package json dependency hierarchy swagger tools tgz root library multer tgz mkdirp tgz x minimist tgz vulnerable library found in head commit a href found in base branches master vulnerability details minimist is vulnerable to prototype pollution via file index js function setkey lines publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimist direct dependency fix resolution ts node fix resolution minimist direct dependency fix resolution express check this box to open an automated fix pr
0
186,884
15,086,800,311
IssuesEvent
2021-02-05 20:57:17
hashicorp/consul-helm
https://api.github.com/repos/hashicorp/consul-helm
closed
Document how to run on Kind
documentation
Kind: https://kind.sigs.k8s.io The only problem is the servers requiring 3 nodes due to their anti-affinity config. Can either run with ```yaml server: replicas: 1 bootstrapExpect: 1 ``` or ```yaml server: affinity: "" ``` Request from https://github.com/hashicorp/consul-helm/issues/322
1.0
Document how to run on Kind - Kind: https://kind.sigs.k8s.io The only problem is the servers requiring 3 nodes due to their anti-affinity config. Can either run with ```yaml server: replicas: 1 bootstrapExpect: 1 ``` or ```yaml server: affinity: "" ``` Request from https://github.com/hashicorp/consul-helm/issues/322
non_process
document how to run on kind kind the only problem is the servers requiring nodes due to their anti affinity config can either run with yaml server replicas bootstrapexpect or yaml server affinity request from
0
415,312
12,127,633,609
IssuesEvent
2020-04-22 19:04:03
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
data race in subchannel code
kind/bug priority/P2
``` ================== WARNING: ThreadSanitizer: data race (pid=17) Read of size 8 at 0x7b5c00000288 by main thread: #0 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::size() const /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:906:24 (liblibgrpc_Uclient_Uchannel.so+0xe4619) #1 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2273:36 (liblibgrpc_Uclient_Uchannel.so+0xe42f0) #2 std::set<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_set.h:604:21 (liblibgrpc_Uclient_Uchannel.so+0xe4248) #3 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:881:34 (liblibgrpc_Uclient_Uchannel.so+0xe1b8e) #4 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:875:24 (liblibgrpc_Uclient_Uchannel.so+0xe1f0f) #5 grpc_core::RefCounted<grpc_core::SubchannelInterface, grpc_core::PolymorphicRefCount>::Unref() /proc/self/cwd/./src/core/lib/gprpp/ref_counted.h:270:7 (liblibgrpc_Ulb_Upolicy_Upick_Ufirst.so+0x224d9) #6 grpc_core::RefCountedPtr<grpc_core::SubchannelInterface>::reset(grpc_core::SubchannelInterface*) /proc/self/cwd/./src/core/lib/gprpp/ref_counted_ptr.h:110:36 (liblibgrpc_Ulb_Upolicy_Upick_Ufirst.so+0x22411) #7 grpc_core::SubchannelData<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::UnrefSubchannelLocked(char const*) /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:291:17 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d6f5) #8 grpc_core::SubchannelData<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::ShutdownLocked() /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:343:3 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d2f2) #9 grpc_core::SubchannelList<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::ShutdownLocked() /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:439:9 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d181) #10 grpc_core::SubchannelList<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::Orphan() /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:196:5 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1761f) #11 void grpc_core::OrphanableDelete::operator()<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList>(grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1608a) #12 std::unique_ptr<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::OrphanableDelete>::reset(grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d9dd) #13 grpc_core::(anonymous namespace)::RoundRobin::ShutdownLocked() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy/round_robin/round_robin.cc:244:20 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x15e79) #14 grpc_core::LoadBalancingPolicy::Orphan() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy.cc:45:3 (liblibgrpc_Uclient_Uchannel.so+0x1266df) #15 void grpc_core::OrphanableDelete::operator()<grpc_core::LoadBalancingPolicy>(grpc_core::LoadBalancingPolicy*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x24a6a) #16 std::unique_ptr<grpc_core::LoadBalancingPolicy, grpc_core::OrphanableDelete>::reset(grpc_core::LoadBalancingPolicy*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x3014d) #17 grpc_core::(anonymous namespace)::GrpcLb::ShutdownLocked() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy/grpclb/grpclb.cc:1408:17 (liblibgrpc_Ulb_Upolicy_Ugrpclb_Usecure.so+0x35976) #18 grpc_core::LoadBalancingPolicy::Orphan() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy.cc:45:3 (liblibgrpc_Uclient_Uchannel.so+0x1266df) #19 void grpc_core::OrphanableDelete::operator()<grpc_core::LoadBalancingPolicy>(grpc_core::LoadBalancingPolicy*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x24a6a) #20 std::unique_ptr<grpc_core::LoadBalancingPolicy, grpc_core::OrphanableDelete>::reset(grpc_core::LoadBalancingPolicy*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x3014d) #21 grpc_core::ResolvingLoadBalancingPolicy::ShutdownLocked() /proc/self/cwd/src/core/ext/filters/client_channel/resolving_lb_policy.cc:217:18 (liblibgrpc_Uclient_Uchannel.so+0x148245) #22 grpc_core::LoadBalancingPolicy::Orphan() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy.cc:45:3 (liblibgrpc_Uclient_Uchannel.so+0x1266df) #23 void grpc_core::OrphanableDelete::operator()<grpc_core::ResolvingLoadBalancingPolicy>(grpc_core::ResolvingLoadBalancingPolicy*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Uclient_Uchannel.so+0xfd4ba) #24 std::unique_ptr<grpc_core::ResolvingLoadBalancingPolicy, grpc_core::OrphanableDelete>::reset(grpc_core::ResolvingLoadBalancingPolicy*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Uclient_Uchannel.so+0xfd2dd) #25 grpc_core::(anonymous namespace)::ChannelData::DestroyResolvingLoadBalancingPolicyLocked() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1592:26 (liblibgrpc_Uclient_Uchannel.so+0xd9948) #26 grpc_core::(anonymous namespace)::ChannelData::StartTransportOpLocked(grpc_transport_op*) /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1830:12 (liblibgrpc_Uclient_Uchannel.so+0xd90c5) #27 grpc_core::(anonymous namespace)::ChannelData::StartTransportOp(grpc_channel_element*, grpc_transport_op*)::$_3::operator()() const /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1869:16 (liblibgrpc_Uclient_Uchannel.so+0xd8a3b) #28 std::_Function_handler<void (), grpc_core::(anonymous namespace)::ChannelData::StartTransportOp(grpc_channel_element*, grpc_transport_op*)::$_3>::_M_invoke(std::_Any_data const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/functional:1871:2 (liblibgrpc_Uclient_Uchannel.so+0xd87fa) #29 std::function<void ()>::operator()() const /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/functional:2267:14 (grpclb_end2end_test+0x54ea36) #30 grpc_core::WorkSerializer::WorkSerializerImpl::Run(std::function<void ()>, grpc_core::DebugLocation const&) /proc/self/cwd/src/core/lib/iomgr/work_serializer.cc:67:5 (liblibgrpc_Ubase_Uc.so+0x15ba31) #31 grpc_core::WorkSerializer::Run(std::function<void ()>, grpc_core::DebugLocation const&) /proc/self/cwd/src/core/lib/iomgr/work_serializer.cc:152:10 (liblibgrpc_Ubase_Uc.so+0x15c0ea) #32 grpc_core::(anonymous namespace)::ChannelData::StartTransportOp(grpc_channel_element*, grpc_transport_op*) /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1868:28 (liblibgrpc_Uclient_Uchannel.so+0xc8b41) #33 grpc_channel_destroy_internal(grpc_channel*) /proc/self/cwd/src/core/lib/surface/channel.cc:514:3 (liblibgrpc_Ubase_Uc.so+0x18caf0) #34 grpc_channel_destroy /proc/self/cwd/src/core/lib/surface/channel.cc:520:3 (liblibgrpc_Ubase_Uc.so+0x18cb99) #35 grpc_impl::Channel::~Channel() /proc/self/cwd/src/cpp/client/channel_cc.cc:61:3 (liblibgrpc++_Ubase.so+0xce10a) #36 std::_Sp_counted_ptr<grpc_impl::Channel*, (__gnu_cxx::_Lock_policy)2>::_M_dispose() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:374:9 (liblibgrpc++_Ubase.so+0xdd828) #37 std::_Sp_counted_base<(__gnu_cxx::_Lock_policy)2>::_M_release() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:150:6 (grpclb_end2end_test+0x55b808) #38 std::__shared_count<(__gnu_cxx::_Lock_policy)2>::~__shared_count() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:659:11 (grpclb_end2end_test+0x55b788) #39 std::__shared_ptr<grpc_impl::Channel, (__gnu_cxx::_Lock_policy)2>::~__shared_ptr() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:925:31 (grpclb_end2end_test+0x55fdb2) #40 std::shared_ptr<grpc_impl::Channel>::~shared_ptr() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr.h:93:11 (grpclb_end2end_test+0x544acb) #41 grpc::testing::(anonymous namespace)::GrpclbEnd2endTest::~GrpclbEnd2endTest() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:368:7 (grpclb_end2end_test+0x506a01) #42 grpc::testing::(anonymous namespace)::UpdatesWithClientLoadReportingTest::~UpdatesWithClientLoadReportingTest() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:1714:7 (grpclb_end2end_test+0x53ae1b) #43 grpc::testing::(anonymous namespace)::UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test::~UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:1719:1 (grpclb_end2end_test+0x537b5b) #44 grpc::testing::(anonymous namespace)::UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test::~UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:1719:1 (grpclb_end2end_test+0x537b9f) #45 testing::Test::DeleteSelf_() /proc/self/cwd/external/com_github_google_googletest/googletest/include/gtest/gtest.h:492:24 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x107a4a) #46 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12443c) #47 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x106b81) #48 testing::TestInfo::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2696:5 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe7fc4) #49 testing::TestSuite::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2822:28 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe8a1a) #50 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:5342:44 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfc2db) #51 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12afdc) #52 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x10af67) #53 testing::UnitTest::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:4930:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfbb0e) #54 RUN_ALL_TESTS() /proc/self/cwd/external/com_github_google_googletest/googletest/include/gtest/gtest.h:2472:46 (grpclb_end2end_test+0x544787) #55 main /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:2054:23 (grpclb_end2end_test+0x505074) Previous write of size 8 at 0x7b5c00000288 by thread T22: #0 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::_M_erase_aux(std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2250:7 (liblibgrpc_Uclient_Uchannel.so+0xe506c) #1 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase[abi:cxx11](std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:1024:2 (liblibgrpc_Uclient_Uchannel.so+0xe4dd2) #2 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::_M_erase_aux(std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2263:4 (liblibgrpc_Uclient_Uchannel.so+0xe4abf) #3 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase[abi:cxx11](std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:1057:2 (liblibgrpc_Uclient_Uchannel.so+0xe46cc) #4 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2274:7 (liblibgrpc_Uclient_Uchannel.so+0xe431f) #5 std::set<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_set.h:604:21 (liblibgrpc_Uclient_Uchannel.so+0xe4248) #6 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:881:34 (liblibgrpc_Uclient_Uchannel.so+0xe1b8e) #7 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:875:24 (liblibgrpc_Uclient_Uchannel.so+0xe1f0f) #8 grpc_core::RefCounted<grpc_core::SubchannelInterface, grpc_core::PolymorphicRefCount>::Unref(grpc_core::DebugLocation const&, char const*) /proc/self/cwd/./src/core/lib/gprpp/ref_counted.h:275:7 (liblibgrpc_Uclient_Uchannel.so+0x106589) #9 grpc_core::RefCountedPtr<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper>::reset(grpc_core::DebugLocation const&, char const*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*) /proc/self/cwd/./src/core/lib/gprpp/ref_counted_ptr.h:115:36 (liblibgrpc_Uclient_Uchannel.so+0xe9ae1) #10 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::WatcherWrapper::~WatcherWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1014:33 (liblibgrpc_Uclient_Uchannel.so+0xe95d2) #11 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::WatcherWrapper::~WatcherWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1014:23 (liblibgrpc_Uclient_Uchannel.so+0xe968f) #12 grpc_core::RefCounted<grpc_core::Subchannel::ConnectivityStateWatcherInterface, grpc_core::PolymorphicRefCount>::Unref() /proc/self/cwd/./src/core/lib/gprpp/ref_counted.h:270:7 (liblibgrpc_Uclient_Uchannel.so+0x106859) #13 grpc_core::RefCountedPtr<grpc_core::Subchannel::ConnectivityStateWatcherInterface>::~RefCountedPtr() /proc/self/cwd/./src/core/lib/gprpp/ref_counted_ptr.h:103:36 (liblibgrpc_Uclient_Uchannel.so+0x1058a5) #14 grpc_core::Subchannel::AsyncWatcherNotifier::~AsyncWatcherNotifier() /proc/self/cwd/src/core/ext/filters/client_channel/subchannel.cc:367:19 (liblibgrpc_Uclient_Uchannel.so+0x16ea02) #15 grpc_core::Subchannel::AsyncWatcherNotifier::AsyncWatcherNotifier(grpc_core::RefCountedPtr<grpc_core::Subchannel::ConnectivityStateWatcherInterface>, grpc_core::Subchannel*, grpc_connectivity_state)::'lambda'(void*, grpc_error*)::operator()(void*, grpc_error*) const /proc/self/cwd/src/core/ext/filters/client_channel/subchannel.cc:377:18 (liblibgrpc_Uclient_Uchannel.so+0x16e97c) #16 grpc_core::Subchannel::AsyncWatcherNotifier::AsyncWatcherNotifier(grpc_core::RefCountedPtr<grpc_core::Subchannel::ConnectivityStateWatcherInterface>, grpc_core::Subchannel*, grpc_connectivity_state)::'lambda'(void*, grpc_error*)::__invoke(void*, grpc_error*) /proc/self/cwd/src/core/ext/filters/client_channel/subchannel.cc:377:18 (liblibgrpc_Uclient_Uchannel.so+0x16e878) #17 exec_ctx_run(grpc_closure*, grpc_error*) /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:40:3 (liblibgrpc_Ubase_Uc.so+0x113437) #18 grpc_core::ExecCtx::Flush() /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:153:9 (liblibgrpc_Ubase_Uc.so+0x11317e) #19 pollset_work(grpc_pollset*, grpc_pollset_worker**, long) /proc/self/cwd/src/core/lib/iomgr/ev_epollex_linux.cc:1136:34 (liblibgrpc_Ubase_Uc.so+0xfbc4b) #20 pollset_work(grpc_pollset*, grpc_pollset_worker**, long) /proc/self/cwd/src/core/lib/iomgr/ev_posix.cc:324:21 (liblibgrpc_Ubase_Uc.so+0x1117d0) #21 grpc_pollset_work(grpc_pollset*, grpc_pollset_worker**, long) /proc/self/cwd/src/core/lib/iomgr/pollset.cc:48:10 (liblibgrpc_Ubase_Uc.so+0x11f66a) #22 run_poller(void*, grpc_error*) /proc/self/cwd/src/core/ext/filters/client_channel/backup_poller.cc:130:7 (liblibgrpc_Uclient_Uchannel.so+0xc4fe3) #23 exec_ctx_run(grpc_closure*, grpc_error*) /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:40:3 (liblibgrpc_Ubase_Uc.so+0x113437) #24 grpc_core::ExecCtx::Flush() /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:153:9 (liblibgrpc_Ubase_Uc.so+0x11317e) #25 run_some_timers() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:134:30 (liblibgrpc_Ubase_Uc.so+0x153d97) #26 timer_main_loop() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:237:9 (liblibgrpc_Ubase_Uc.so+0x1539b3) #27 timer_thread(void*) /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:284:3 (liblibgrpc_Ubase_Uc.so+0x1538aa) #28 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::operator()(void*) const /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:140:27 (liblibgpr_Ubase.so+0x160fe) #29 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::__invoke(void*) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:110:25 (liblibgpr_Ubase.so+0x15ef8) Location is heap block of size 856 at 0x7b5c00000000 allocated by main thread: #0 calloc /tmp/clang-build/src/compiler-rt/lib/tsan/rtl/tsan_interceptors.cpp:667:5 (grpclb_end2end_test+0x492cc0) #1 gpr_zalloc /proc/self/cwd/src/core/lib/gpr/alloc.cc:43:7 (liblibgpr_Ubase.so+0x8b2a) #2 grpc_channel_stack_builder_finish(grpc_channel_stack_builder*, unsigned long, int, void (*)(void*, grpc_error*), void*, void**) /proc/self/cwd/src/core/lib/channel/channel_stack_builder.cc:292:13 (liblibgrpc_Ubase_Uc.so+0x9a8d0) #3 grpc_channel_create_with_builder(grpc_channel_stack_builder*, grpc_channel_stack_type) /proc/self/cwd/src/core/lib/surface/channel.cc:79:23 (liblibgrpc_Ubase_Uc.so+0x189975) #4 grpc_channel_create(char const*, grpc_channel_args const*, grpc_channel_stack_type, grpc_transport*, grpc_resource_user*) /proc/self/cwd/src/core/lib/surface/channel.cc:290:7 (liblibgrpc_Ubase_Uc.so+0x18aaae) #5 grpc_core::(anonymous namespace)::CreateChannel(char const*, grpc_channel_args const*) /proc/self/cwd/src/core/ext/transport/chttp2/client/secure/secure_channel_create.cc:169:7 (liblibgrpc_Utransport_Uchttp2_Uclient_Usecure.so+0x6b73) #6 grpc_secure_channel_create /proc/self/cwd/src/core/ext/transport/chttp2/client/secure/secure_channel_create.cc:217:15 (liblibgrpc_Utransport_Uchttp2_Uclient_Usecure.so+0x68bb) #7 grpc_impl::SecureChannelCredentials::CreateChannelWithInterceptors(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, grpc_impl::ChannelArguments const&, std::vector<std::unique_ptr<grpc::experimental::ClientInterceptorFactoryInterface, std::default_delete<grpc::experimental::ClientInterceptorFactoryInterface> >, std::allocator<std::unique_ptr<grpc::experimental::ClientInterceptorFactoryInterface, std::default_delete<grpc::experimental::ClientInterceptorFactoryInterface> > > >) /proc/self/cwd/src/cpp/client/secure_credentials.cc:67:7 (liblibgrpc++.so+0x323d6) #8 grpc_impl::SecureChannelCredentials::CreateChannelImpl(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, grpc_impl::ChannelArguments const&) /proc/self/cwd/src/cpp/client/secure_credentials.cc:51:10 (liblibgrpc++.so+0x32287) #9 grpc_impl::CreateCustomChannelImpl(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::shared_ptr<grpc_impl::ChannelCredentials> const&, grpc_impl::ChannelArguments const&) /proc/self/cwd/src/cpp/client/create_channel.cc:42:25 (liblibgrpc++_Ubase.so+0xdc6da) #10 grpc::CreateCustomChannel(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::shared_ptr<grpc_impl::ChannelCredentials> const&, grpc_impl::ChannelArguments const&) /proc/self/cwd/include/grpcpp/create_channel.h:37:10 (grpclb_end2end_test+0x514b3e) #11 grpc::testing::(anonymous namespace)::GrpclbEnd2endTest::ResetStub(int, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&) /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:447:16 (grpclb_end2end_test+0x50ae04) #12 grpc::testing::(anonymous namespace)::GrpclbEnd2endTest::SetUp() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:405:5 (grpclb_end2end_test+0x50585f) #13 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12443c) #14 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x106b81) #15 testing::Test::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2509:3 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe6dfa) #16 testing::TestInfo::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2690:11 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe7f4b) #17 testing::TestSuite::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2822:28 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe8a1a) #18 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:5342:44 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfc2db) #19 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12afdc) #20 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x10af67) #21 testing::UnitTest::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:4930:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfbb0e) #22 RUN_ALL_TESTS() /proc/self/cwd/external/com_github_google_googletest/googletest/include/gtest/gtest.h:2472:46 (grpclb_end2end_test+0x544787) #23 main /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:2054:23 (grpclb_end2end_test+0x505074) Thread T22 'grpc_global_tim' (tid=575, running) created by thread T37 at: #0 pthread_create /tmp/clang-build/src/compiler-rt/lib/tsan/rtl/tsan_interceptors.cpp:967:3 (grpclb_end2end_test+0x49432b) #1 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:109:10 (liblibgpr_Ubase.so+0x15a60) #2 grpc_core::Thread::Thread(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:186:15 (liblibgpr_Ubase.so+0x15326) #3 start_timer_thread_and_unlock() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:92:13 (liblibgrpc_Ubase_Uc.so+0x1537e3) #4 run_some_timers() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:118:5 (liblibgrpc_Ubase_Uc.so+0x153c82) #5 timer_main_loop() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:237:9 (liblibgrpc_Ubase_Uc.so+0x1539b3) #6 timer_thread(void*) /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:284:3 (liblibgrpc_Ubase_Uc.so+0x1538aa) #7 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::operator()(void*) const /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:140:27 (liblibgpr_Ubase.so+0x160fe) #8 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::__invoke(void*) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:110:25 (liblibgpr_Ubase.so+0x15ef8) SUMMARY: ThreadSanitizer: data race /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:906:24 in std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::size() const ================== ``` Seen on #21361 https://source.cloud.google.com/results/invocations/e6109732-35d9-4e2b-a1aa-1b77f7d2993d/targets/%2F%2Ftest%2Fcpp%2Fend2end:grpclb_end2end_test@poller%3Depollex/log
1.0
data race in subchannel code - ``` ================== WARNING: ThreadSanitizer: data race (pid=17) Read of size 8 at 0x7b5c00000288 by main thread: #0 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::size() const /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:906:24 (liblibgrpc_Uclient_Uchannel.so+0xe4619) #1 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2273:36 (liblibgrpc_Uclient_Uchannel.so+0xe42f0) #2 std::set<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_set.h:604:21 (liblibgrpc_Uclient_Uchannel.so+0xe4248) #3 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:881:34 (liblibgrpc_Uclient_Uchannel.so+0xe1b8e) #4 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:875:24 (liblibgrpc_Uclient_Uchannel.so+0xe1f0f) #5 grpc_core::RefCounted<grpc_core::SubchannelInterface, grpc_core::PolymorphicRefCount>::Unref() /proc/self/cwd/./src/core/lib/gprpp/ref_counted.h:270:7 (liblibgrpc_Ulb_Upolicy_Upick_Ufirst.so+0x224d9) #6 grpc_core::RefCountedPtr<grpc_core::SubchannelInterface>::reset(grpc_core::SubchannelInterface*) /proc/self/cwd/./src/core/lib/gprpp/ref_counted_ptr.h:110:36 (liblibgrpc_Ulb_Upolicy_Upick_Ufirst.so+0x22411) #7 grpc_core::SubchannelData<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::UnrefSubchannelLocked(char const*) /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:291:17 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d6f5) #8 grpc_core::SubchannelData<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::ShutdownLocked() /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:343:3 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d2f2) #9 grpc_core::SubchannelList<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::ShutdownLocked() /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:439:9 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d181) #10 grpc_core::SubchannelList<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelData>::Orphan() /proc/self/cwd/./src/core/ext/filters/client_channel/lb_policy/subchannel_list.h:196:5 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1761f) #11 void grpc_core::OrphanableDelete::operator()<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList>(grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1608a) #12 std::unique_ptr<grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList, grpc_core::OrphanableDelete>::reset(grpc_core::(anonymous namespace)::RoundRobin::RoundRobinSubchannelList*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x1d9dd) #13 grpc_core::(anonymous namespace)::RoundRobin::ShutdownLocked() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy/round_robin/round_robin.cc:244:20 (liblibgrpc_Ulb_Upolicy_Uround_Urobin.so+0x15e79) #14 grpc_core::LoadBalancingPolicy::Orphan() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy.cc:45:3 (liblibgrpc_Uclient_Uchannel.so+0x1266df) #15 void grpc_core::OrphanableDelete::operator()<grpc_core::LoadBalancingPolicy>(grpc_core::LoadBalancingPolicy*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x24a6a) #16 std::unique_ptr<grpc_core::LoadBalancingPolicy, grpc_core::OrphanableDelete>::reset(grpc_core::LoadBalancingPolicy*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x3014d) #17 grpc_core::(anonymous namespace)::GrpcLb::ShutdownLocked() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy/grpclb/grpclb.cc:1408:17 (liblibgrpc_Ulb_Upolicy_Ugrpclb_Usecure.so+0x35976) #18 grpc_core::LoadBalancingPolicy::Orphan() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy.cc:45:3 (liblibgrpc_Uclient_Uchannel.so+0x1266df) #19 void grpc_core::OrphanableDelete::operator()<grpc_core::LoadBalancingPolicy>(grpc_core::LoadBalancingPolicy*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x24a6a) #20 std::unique_ptr<grpc_core::LoadBalancingPolicy, grpc_core::OrphanableDelete>::reset(grpc_core::LoadBalancingPolicy*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Ulb_Upolicy_Ucds_Usecure.so+0x3014d) #21 grpc_core::ResolvingLoadBalancingPolicy::ShutdownLocked() /proc/self/cwd/src/core/ext/filters/client_channel/resolving_lb_policy.cc:217:18 (liblibgrpc_Uclient_Uchannel.so+0x148245) #22 grpc_core::LoadBalancingPolicy::Orphan() /proc/self/cwd/src/core/ext/filters/client_channel/lb_policy.cc:45:3 (liblibgrpc_Uclient_Uchannel.so+0x1266df) #23 void grpc_core::OrphanableDelete::operator()<grpc_core::ResolvingLoadBalancingPolicy>(grpc_core::ResolvingLoadBalancingPolicy*) /proc/self/cwd/./src/core/lib/gprpp/orphanable.h:63:8 (liblibgrpc_Uclient_Uchannel.so+0xfd4ba) #24 std::unique_ptr<grpc_core::ResolvingLoadBalancingPolicy, grpc_core::OrphanableDelete>::reset(grpc_core::ResolvingLoadBalancingPolicy*) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/unique_ptr.h:344:4 (liblibgrpc_Uclient_Uchannel.so+0xfd2dd) #25 grpc_core::(anonymous namespace)::ChannelData::DestroyResolvingLoadBalancingPolicyLocked() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1592:26 (liblibgrpc_Uclient_Uchannel.so+0xd9948) #26 grpc_core::(anonymous namespace)::ChannelData::StartTransportOpLocked(grpc_transport_op*) /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1830:12 (liblibgrpc_Uclient_Uchannel.so+0xd90c5) #27 grpc_core::(anonymous namespace)::ChannelData::StartTransportOp(grpc_channel_element*, grpc_transport_op*)::$_3::operator()() const /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1869:16 (liblibgrpc_Uclient_Uchannel.so+0xd8a3b) #28 std::_Function_handler<void (), grpc_core::(anonymous namespace)::ChannelData::StartTransportOp(grpc_channel_element*, grpc_transport_op*)::$_3>::_M_invoke(std::_Any_data const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/functional:1871:2 (liblibgrpc_Uclient_Uchannel.so+0xd87fa) #29 std::function<void ()>::operator()() const /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/functional:2267:14 (grpclb_end2end_test+0x54ea36) #30 grpc_core::WorkSerializer::WorkSerializerImpl::Run(std::function<void ()>, grpc_core::DebugLocation const&) /proc/self/cwd/src/core/lib/iomgr/work_serializer.cc:67:5 (liblibgrpc_Ubase_Uc.so+0x15ba31) #31 grpc_core::WorkSerializer::Run(std::function<void ()>, grpc_core::DebugLocation const&) /proc/self/cwd/src/core/lib/iomgr/work_serializer.cc:152:10 (liblibgrpc_Ubase_Uc.so+0x15c0ea) #32 grpc_core::(anonymous namespace)::ChannelData::StartTransportOp(grpc_channel_element*, grpc_transport_op*) /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1868:28 (liblibgrpc_Uclient_Uchannel.so+0xc8b41) #33 grpc_channel_destroy_internal(grpc_channel*) /proc/self/cwd/src/core/lib/surface/channel.cc:514:3 (liblibgrpc_Ubase_Uc.so+0x18caf0) #34 grpc_channel_destroy /proc/self/cwd/src/core/lib/surface/channel.cc:520:3 (liblibgrpc_Ubase_Uc.so+0x18cb99) #35 grpc_impl::Channel::~Channel() /proc/self/cwd/src/cpp/client/channel_cc.cc:61:3 (liblibgrpc++_Ubase.so+0xce10a) #36 std::_Sp_counted_ptr<grpc_impl::Channel*, (__gnu_cxx::_Lock_policy)2>::_M_dispose() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:374:9 (liblibgrpc++_Ubase.so+0xdd828) #37 std::_Sp_counted_base<(__gnu_cxx::_Lock_policy)2>::_M_release() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:150:6 (grpclb_end2end_test+0x55b808) #38 std::__shared_count<(__gnu_cxx::_Lock_policy)2>::~__shared_count() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:659:11 (grpclb_end2end_test+0x55b788) #39 std::__shared_ptr<grpc_impl::Channel, (__gnu_cxx::_Lock_policy)2>::~__shared_ptr() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr_base.h:925:31 (grpclb_end2end_test+0x55fdb2) #40 std::shared_ptr<grpc_impl::Channel>::~shared_ptr() /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/shared_ptr.h:93:11 (grpclb_end2end_test+0x544acb) #41 grpc::testing::(anonymous namespace)::GrpclbEnd2endTest::~GrpclbEnd2endTest() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:368:7 (grpclb_end2end_test+0x506a01) #42 grpc::testing::(anonymous namespace)::UpdatesWithClientLoadReportingTest::~UpdatesWithClientLoadReportingTest() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:1714:7 (grpclb_end2end_test+0x53ae1b) #43 grpc::testing::(anonymous namespace)::UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test::~UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:1719:1 (grpclb_end2end_test+0x537b5b) #44 grpc::testing::(anonymous namespace)::UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test::~UpdatesWithClientLoadReportingTest_ReresolveDeadBalancer_Test() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:1719:1 (grpclb_end2end_test+0x537b9f) #45 testing::Test::DeleteSelf_() /proc/self/cwd/external/com_github_google_googletest/googletest/include/gtest/gtest.h:492:24 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x107a4a) #46 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12443c) #47 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x106b81) #48 testing::TestInfo::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2696:5 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe7fc4) #49 testing::TestSuite::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2822:28 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe8a1a) #50 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:5342:44 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfc2db) #51 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12afdc) #52 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x10af67) #53 testing::UnitTest::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:4930:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfbb0e) #54 RUN_ALL_TESTS() /proc/self/cwd/external/com_github_google_googletest/googletest/include/gtest/gtest.h:2472:46 (grpclb_end2end_test+0x544787) #55 main /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:2054:23 (grpclb_end2end_test+0x505074) Previous write of size 8 at 0x7b5c00000288 by thread T22: #0 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::_M_erase_aux(std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2250:7 (liblibgrpc_Uclient_Uchannel.so+0xe506c) #1 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase[abi:cxx11](std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:1024:2 (liblibgrpc_Uclient_Uchannel.so+0xe4dd2) #2 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::_M_erase_aux(std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2263:4 (liblibgrpc_Uclient_Uchannel.so+0xe4abf) #3 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase[abi:cxx11](std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::_Rb_tree_const_iterator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:1057:2 (liblibgrpc_Uclient_Uchannel.so+0xe46cc) #4 std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:2274:7 (liblibgrpc_Uclient_Uchannel.so+0xe431f) #5 std::set<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::erase(grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper* const&) /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_set.h:604:21 (liblibgrpc_Uclient_Uchannel.so+0xe4248) #6 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:881:34 (liblibgrpc_Uclient_Uchannel.so+0xe1b8e) #7 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::~SubchannelWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:875:24 (liblibgrpc_Uclient_Uchannel.so+0xe1f0f) #8 grpc_core::RefCounted<grpc_core::SubchannelInterface, grpc_core::PolymorphicRefCount>::Unref(grpc_core::DebugLocation const&, char const*) /proc/self/cwd/./src/core/lib/gprpp/ref_counted.h:275:7 (liblibgrpc_Uclient_Uchannel.so+0x106589) #9 grpc_core::RefCountedPtr<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper>::reset(grpc_core::DebugLocation const&, char const*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*) /proc/self/cwd/./src/core/lib/gprpp/ref_counted_ptr.h:115:36 (liblibgrpc_Uclient_Uchannel.so+0xe9ae1) #10 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::WatcherWrapper::~WatcherWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1014:33 (liblibgrpc_Uclient_Uchannel.so+0xe95d2) #11 grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper::WatcherWrapper::~WatcherWrapper() /proc/self/cwd/src/core/ext/filters/client_channel/client_channel.cc:1014:23 (liblibgrpc_Uclient_Uchannel.so+0xe968f) #12 grpc_core::RefCounted<grpc_core::Subchannel::ConnectivityStateWatcherInterface, grpc_core::PolymorphicRefCount>::Unref() /proc/self/cwd/./src/core/lib/gprpp/ref_counted.h:270:7 (liblibgrpc_Uclient_Uchannel.so+0x106859) #13 grpc_core::RefCountedPtr<grpc_core::Subchannel::ConnectivityStateWatcherInterface>::~RefCountedPtr() /proc/self/cwd/./src/core/lib/gprpp/ref_counted_ptr.h:103:36 (liblibgrpc_Uclient_Uchannel.so+0x1058a5) #14 grpc_core::Subchannel::AsyncWatcherNotifier::~AsyncWatcherNotifier() /proc/self/cwd/src/core/ext/filters/client_channel/subchannel.cc:367:19 (liblibgrpc_Uclient_Uchannel.so+0x16ea02) #15 grpc_core::Subchannel::AsyncWatcherNotifier::AsyncWatcherNotifier(grpc_core::RefCountedPtr<grpc_core::Subchannel::ConnectivityStateWatcherInterface>, grpc_core::Subchannel*, grpc_connectivity_state)::'lambda'(void*, grpc_error*)::operator()(void*, grpc_error*) const /proc/self/cwd/src/core/ext/filters/client_channel/subchannel.cc:377:18 (liblibgrpc_Uclient_Uchannel.so+0x16e97c) #16 grpc_core::Subchannel::AsyncWatcherNotifier::AsyncWatcherNotifier(grpc_core::RefCountedPtr<grpc_core::Subchannel::ConnectivityStateWatcherInterface>, grpc_core::Subchannel*, grpc_connectivity_state)::'lambda'(void*, grpc_error*)::__invoke(void*, grpc_error*) /proc/self/cwd/src/core/ext/filters/client_channel/subchannel.cc:377:18 (liblibgrpc_Uclient_Uchannel.so+0x16e878) #17 exec_ctx_run(grpc_closure*, grpc_error*) /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:40:3 (liblibgrpc_Ubase_Uc.so+0x113437) #18 grpc_core::ExecCtx::Flush() /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:153:9 (liblibgrpc_Ubase_Uc.so+0x11317e) #19 pollset_work(grpc_pollset*, grpc_pollset_worker**, long) /proc/self/cwd/src/core/lib/iomgr/ev_epollex_linux.cc:1136:34 (liblibgrpc_Ubase_Uc.so+0xfbc4b) #20 pollset_work(grpc_pollset*, grpc_pollset_worker**, long) /proc/self/cwd/src/core/lib/iomgr/ev_posix.cc:324:21 (liblibgrpc_Ubase_Uc.so+0x1117d0) #21 grpc_pollset_work(grpc_pollset*, grpc_pollset_worker**, long) /proc/self/cwd/src/core/lib/iomgr/pollset.cc:48:10 (liblibgrpc_Ubase_Uc.so+0x11f66a) #22 run_poller(void*, grpc_error*) /proc/self/cwd/src/core/ext/filters/client_channel/backup_poller.cc:130:7 (liblibgrpc_Uclient_Uchannel.so+0xc4fe3) #23 exec_ctx_run(grpc_closure*, grpc_error*) /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:40:3 (liblibgrpc_Ubase_Uc.so+0x113437) #24 grpc_core::ExecCtx::Flush() /proc/self/cwd/src/core/lib/iomgr/exec_ctx.cc:153:9 (liblibgrpc_Ubase_Uc.so+0x11317e) #25 run_some_timers() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:134:30 (liblibgrpc_Ubase_Uc.so+0x153d97) #26 timer_main_loop() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:237:9 (liblibgrpc_Ubase_Uc.so+0x1539b3) #27 timer_thread(void*) /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:284:3 (liblibgrpc_Ubase_Uc.so+0x1538aa) #28 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::operator()(void*) const /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:140:27 (liblibgpr_Ubase.so+0x160fe) #29 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::__invoke(void*) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:110:25 (liblibgpr_Ubase.so+0x15ef8) Location is heap block of size 856 at 0x7b5c00000000 allocated by main thread: #0 calloc /tmp/clang-build/src/compiler-rt/lib/tsan/rtl/tsan_interceptors.cpp:667:5 (grpclb_end2end_test+0x492cc0) #1 gpr_zalloc /proc/self/cwd/src/core/lib/gpr/alloc.cc:43:7 (liblibgpr_Ubase.so+0x8b2a) #2 grpc_channel_stack_builder_finish(grpc_channel_stack_builder*, unsigned long, int, void (*)(void*, grpc_error*), void*, void**) /proc/self/cwd/src/core/lib/channel/channel_stack_builder.cc:292:13 (liblibgrpc_Ubase_Uc.so+0x9a8d0) #3 grpc_channel_create_with_builder(grpc_channel_stack_builder*, grpc_channel_stack_type) /proc/self/cwd/src/core/lib/surface/channel.cc:79:23 (liblibgrpc_Ubase_Uc.so+0x189975) #4 grpc_channel_create(char const*, grpc_channel_args const*, grpc_channel_stack_type, grpc_transport*, grpc_resource_user*) /proc/self/cwd/src/core/lib/surface/channel.cc:290:7 (liblibgrpc_Ubase_Uc.so+0x18aaae) #5 grpc_core::(anonymous namespace)::CreateChannel(char const*, grpc_channel_args const*) /proc/self/cwd/src/core/ext/transport/chttp2/client/secure/secure_channel_create.cc:169:7 (liblibgrpc_Utransport_Uchttp2_Uclient_Usecure.so+0x6b73) #6 grpc_secure_channel_create /proc/self/cwd/src/core/ext/transport/chttp2/client/secure/secure_channel_create.cc:217:15 (liblibgrpc_Utransport_Uchttp2_Uclient_Usecure.so+0x68bb) #7 grpc_impl::SecureChannelCredentials::CreateChannelWithInterceptors(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, grpc_impl::ChannelArguments const&, std::vector<std::unique_ptr<grpc::experimental::ClientInterceptorFactoryInterface, std::default_delete<grpc::experimental::ClientInterceptorFactoryInterface> >, std::allocator<std::unique_ptr<grpc::experimental::ClientInterceptorFactoryInterface, std::default_delete<grpc::experimental::ClientInterceptorFactoryInterface> > > >) /proc/self/cwd/src/cpp/client/secure_credentials.cc:67:7 (liblibgrpc++.so+0x323d6) #8 grpc_impl::SecureChannelCredentials::CreateChannelImpl(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, grpc_impl::ChannelArguments const&) /proc/self/cwd/src/cpp/client/secure_credentials.cc:51:10 (liblibgrpc++.so+0x32287) #9 grpc_impl::CreateCustomChannelImpl(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::shared_ptr<grpc_impl::ChannelCredentials> const&, grpc_impl::ChannelArguments const&) /proc/self/cwd/src/cpp/client/create_channel.cc:42:25 (liblibgrpc++_Ubase.so+0xdc6da) #10 grpc::CreateCustomChannel(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, std::shared_ptr<grpc_impl::ChannelCredentials> const&, grpc_impl::ChannelArguments const&) /proc/self/cwd/include/grpcpp/create_channel.h:37:10 (grpclb_end2end_test+0x514b3e) #11 grpc::testing::(anonymous namespace)::GrpclbEnd2endTest::ResetStub(int, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&) /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:447:16 (grpclb_end2end_test+0x50ae04) #12 grpc::testing::(anonymous namespace)::GrpclbEnd2endTest::SetUp() /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:405:5 (grpclb_end2end_test+0x50585f) #13 void testing::internal::HandleSehExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12443c) #14 void testing::internal::HandleExceptionsInMethodIfSupported<testing::Test, void>(testing::Test*, void (testing::Test::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x106b81) #15 testing::Test::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2509:3 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe6dfa) #16 testing::TestInfo::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2690:11 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe7f4b) #17 testing::TestSuite::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2822:28 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xe8a1a) #18 testing::internal::UnitTestImpl::RunAllTests() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:5342:44 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfc2db) #19 bool testing::internal::HandleSehExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2439:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x12afdc) #20 bool testing::internal::HandleExceptionsInMethodIfSupported<testing::internal::UnitTestImpl, bool>(testing::internal::UnitTestImpl*, bool (testing::internal::UnitTestImpl::*)(), char const*) /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:2475:14 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0x10af67) #21 testing::UnitTest::Run() /proc/self/cwd/external/com_github_google_googletest/googletest/src/gtest.cc:4930:10 (libexternal_Scom_Ugithub_Ugoogle_Ugoogletest_Slibgtest.so+0xfbb0e) #22 RUN_ALL_TESTS() /proc/self/cwd/external/com_github_google_googletest/googletest/include/gtest/gtest.h:2472:46 (grpclb_end2end_test+0x544787) #23 main /proc/self/cwd/test/cpp/end2end/grpclb_end2end_test.cc:2054:23 (grpclb_end2end_test+0x505074) Thread T22 'grpc_global_tim' (tid=575, running) created by thread T37 at: #0 pthread_create /tmp/clang-build/src/compiler-rt/lib/tsan/rtl/tsan_interceptors.cpp:967:3 (grpclb_end2end_test+0x49432b) #1 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:109:10 (liblibgpr_Ubase.so+0x15a60) #2 grpc_core::Thread::Thread(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:186:15 (liblibgpr_Ubase.so+0x15326) #3 start_timer_thread_and_unlock() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:92:13 (liblibgrpc_Ubase_Uc.so+0x1537e3) #4 run_some_timers() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:118:5 (liblibgrpc_Ubase_Uc.so+0x153c82) #5 timer_main_loop() /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:237:9 (liblibgrpc_Ubase_Uc.so+0x1539b3) #6 timer_thread(void*) /proc/self/cwd/src/core/lib/iomgr/timer_manager.cc:284:3 (liblibgrpc_Ubase_Uc.so+0x1538aa) #7 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::operator()(void*) const /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:140:27 (liblibgpr_Ubase.so+0x160fe) #8 grpc_core::(anonymous namespace)::ThreadInternalsPosix::ThreadInternalsPosix(char const*, void (*)(void*), void*, bool*, grpc_core::Thread::Options const&)::'lambda'(void*)::__invoke(void*) /proc/self/cwd/src/core/lib/gprpp/thd_posix.cc:110:25 (liblibgpr_Ubase.so+0x15ef8) SUMMARY: ThreadSanitizer: data race /usr/lib/gcc/x86_64-linux-gnu/5.4.0/../../../../include/c++/5.4.0/bits/stl_tree.h:906:24 in std::_Rb_tree<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*, std::_Identity<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::less<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*>, std::allocator<grpc_core::(anonymous namespace)::ChannelData::SubchannelWrapper*> >::size() const ================== ``` Seen on #21361 https://source.cloud.google.com/results/invocations/e6109732-35d9-4e2b-a1aa-1b77f7d2993d/targets/%2F%2Ftest%2Fcpp%2Fend2end:grpclb_end2end_test@poller%3Depollex/log
non_process
data race in subchannel code warning threadsanitizer data race pid read of size at by main thread std rb tree std less std allocator size const usr lib gcc linux gnu include c bits stl tree h liblibgrpc uclient uchannel so std rb tree std less std allocator erase grpc core anonymous namespace channeldata subchannelwrapper const usr lib gcc linux gnu include c bits stl tree h liblibgrpc uclient uchannel so std set std allocator erase grpc core anonymous namespace channeldata subchannelwrapper const usr lib gcc linux gnu include c bits stl set h liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata subchannelwrapper subchannelwrapper proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata subchannelwrapper subchannelwrapper proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core refcounted unref proc self cwd src core lib gprpp ref counted h liblibgrpc ulb upolicy upick ufirst so grpc core refcountedptr reset grpc core subchannelinterface proc self cwd src core lib gprpp ref counted ptr h liblibgrpc ulb upolicy upick ufirst so grpc core subchanneldata unrefsubchannellocked char const proc self cwd src core ext filters client channel lb policy subchannel list h liblibgrpc ulb upolicy uround urobin so grpc core subchanneldata shutdownlocked proc self cwd src core ext filters client channel lb policy subchannel list h liblibgrpc ulb upolicy uround urobin so grpc core subchannellist shutdownlocked proc self cwd src core ext filters client channel lb policy subchannel list h liblibgrpc ulb upolicy uround urobin so grpc core subchannellist orphan proc self cwd src core ext filters client channel lb policy subchannel list h liblibgrpc ulb upolicy uround urobin so void grpc core orphanabledelete operator grpc core anonymous namespace roundrobin roundrobinsubchannellist proc self cwd src core lib gprpp orphanable h liblibgrpc ulb upolicy uround urobin so std unique ptr reset grpc core anonymous namespace roundrobin roundrobinsubchannellist usr lib gcc linux gnu include c bits unique ptr h liblibgrpc ulb upolicy uround urobin so grpc core anonymous namespace roundrobin shutdownlocked proc self cwd src core ext filters client channel lb policy round robin round robin cc liblibgrpc ulb upolicy uround urobin so grpc core loadbalancingpolicy orphan proc self cwd src core ext filters client channel lb policy cc liblibgrpc uclient uchannel so void grpc core orphanabledelete operator grpc core loadbalancingpolicy proc self cwd src core lib gprpp orphanable h liblibgrpc ulb upolicy ucds usecure so std unique ptr reset grpc core loadbalancingpolicy usr lib gcc linux gnu include c bits unique ptr h liblibgrpc ulb upolicy ucds usecure so grpc core anonymous namespace grpclb shutdownlocked proc self cwd src core ext filters client channel lb policy grpclb grpclb cc liblibgrpc ulb upolicy ugrpclb usecure so grpc core loadbalancingpolicy orphan proc self cwd src core ext filters client channel lb policy cc liblibgrpc uclient uchannel so void grpc core orphanabledelete operator grpc core loadbalancingpolicy proc self cwd src core lib gprpp orphanable h liblibgrpc ulb upolicy ucds usecure so std unique ptr reset grpc core loadbalancingpolicy usr lib gcc linux gnu include c bits unique ptr h liblibgrpc ulb upolicy ucds usecure so grpc core resolvingloadbalancingpolicy shutdownlocked proc self cwd src core ext filters client channel resolving lb policy cc liblibgrpc uclient uchannel so grpc core loadbalancingpolicy orphan proc self cwd src core ext filters client channel lb policy cc liblibgrpc uclient uchannel so void grpc core orphanabledelete operator grpc core resolvingloadbalancingpolicy proc self cwd src core lib gprpp orphanable h liblibgrpc uclient uchannel so std unique ptr reset grpc core resolvingloadbalancingpolicy usr lib gcc linux gnu include c bits unique ptr h liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata destroyresolvingloadbalancingpolicylocked proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata starttransportoplocked grpc transport op proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata starttransportop grpc channel element grpc transport op operator const proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so std function handler m invoke std any data const usr lib gcc linux gnu include c functional liblibgrpc uclient uchannel so std function operator const usr lib gcc linux gnu include c functional grpclb test grpc core workserializer workserializerimpl run std function grpc core debuglocation const proc self cwd src core lib iomgr work serializer cc liblibgrpc ubase uc so grpc core workserializer run std function grpc core debuglocation const proc self cwd src core lib iomgr work serializer cc liblibgrpc ubase uc so grpc core anonymous namespace channeldata starttransportop grpc channel element grpc transport op proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc channel destroy internal grpc channel proc self cwd src core lib surface channel cc liblibgrpc ubase uc so grpc channel destroy proc self cwd src core lib surface channel cc liblibgrpc ubase uc so grpc impl channel channel proc self cwd src cpp client channel cc cc liblibgrpc ubase so std sp counted ptr m dispose usr lib gcc linux gnu include c bits shared ptr base h liblibgrpc ubase so std sp counted base m release usr lib gcc linux gnu include c bits shared ptr base h grpclb test std shared count shared count usr lib gcc linux gnu include c bits shared ptr base h grpclb test std shared ptr shared ptr usr lib gcc linux gnu include c bits shared ptr base h grpclb test std shared ptr shared ptr usr lib gcc linux gnu include c bits shared ptr h grpclb test grpc testing anonymous namespace proc self cwd test cpp grpclb test cc grpclb test grpc testing anonymous namespace updateswithclientloadreportingtest updateswithclientloadreportingtest proc self cwd test cpp grpclb test cc grpclb test grpc testing anonymous namespace updateswithclientloadreportingtest reresolvedeadbalancer test updateswithclientloadreportingtest reresolvedeadbalancer test proc self cwd test cpp grpclb test cc grpclb test grpc testing anonymous namespace updateswithclientloadreportingtest reresolvedeadbalancer test updateswithclientloadreportingtest reresolvedeadbalancer test proc self cwd test cpp grpclb test cc grpclb test testing test deleteself proc self cwd external com github google googletest googletest include gtest gtest h libexternal scom ugithub ugoogle ugoogletest slibgtest so void testing internal handlesehexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so void testing internal handleexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing testinfo run proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing testsuite run proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing internal unittestimpl runalltests proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so bool testing internal handlesehexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing unittest run proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so run all tests proc self cwd external com github google googletest googletest include gtest gtest h grpclb test main proc self cwd test cpp grpclb test cc grpclb test previous write of size at by thread std rb tree std less std allocator m erase aux std rb tree const iterator usr lib gcc linux gnu include c bits stl tree h liblibgrpc uclient uchannel so std rb tree std less std allocator erase std rb tree const iterator usr lib gcc linux gnu include c bits stl tree h liblibgrpc uclient uchannel so std rb tree std less std allocator m erase aux std rb tree const iterator std rb tree const iterator usr lib gcc linux gnu include c bits stl tree h liblibgrpc uclient uchannel so std rb tree std less std allocator erase std rb tree const iterator std rb tree const iterator usr lib gcc linux gnu include c bits stl tree h liblibgrpc uclient uchannel so std rb tree std less std allocator erase grpc core anonymous namespace channeldata subchannelwrapper const usr lib gcc linux gnu include c bits stl tree h liblibgrpc uclient uchannel so std set std allocator erase grpc core anonymous namespace channeldata subchannelwrapper const usr lib gcc linux gnu include c bits stl set h liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata subchannelwrapper subchannelwrapper proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata subchannelwrapper subchannelwrapper proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core refcounted unref grpc core debuglocation const char const proc self cwd src core lib gprpp ref counted h liblibgrpc uclient uchannel so grpc core refcountedptr reset grpc core debuglocation const char const grpc core anonymous namespace channeldata subchannelwrapper proc self cwd src core lib gprpp ref counted ptr h liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata subchannelwrapper watcherwrapper watcherwrapper proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core anonymous namespace channeldata subchannelwrapper watcherwrapper watcherwrapper proc self cwd src core ext filters client channel client channel cc liblibgrpc uclient uchannel so grpc core refcounted unref proc self cwd src core lib gprpp ref counted h liblibgrpc uclient uchannel so grpc core refcountedptr refcountedptr proc self cwd src core lib gprpp ref counted ptr h liblibgrpc uclient uchannel so grpc core subchannel asyncwatchernotifier asyncwatchernotifier proc self cwd src core ext filters client channel subchannel cc liblibgrpc uclient uchannel so grpc core subchannel asyncwatchernotifier asyncwatchernotifier grpc core refcountedptr grpc core subchannel grpc connectivity state lambda void grpc error operator void grpc error const proc self cwd src core ext filters client channel subchannel cc liblibgrpc uclient uchannel so grpc core subchannel asyncwatchernotifier asyncwatchernotifier grpc core refcountedptr grpc core subchannel grpc connectivity state lambda void grpc error invoke void grpc error proc self cwd src core ext filters client channel subchannel cc liblibgrpc uclient uchannel so exec ctx run grpc closure grpc error proc self cwd src core lib iomgr exec ctx cc liblibgrpc ubase uc so grpc core execctx flush proc self cwd src core lib iomgr exec ctx cc liblibgrpc ubase uc so pollset work grpc pollset grpc pollset worker long proc self cwd src core lib iomgr ev epollex linux cc liblibgrpc ubase uc so pollset work grpc pollset grpc pollset worker long proc self cwd src core lib iomgr ev posix cc liblibgrpc ubase uc so grpc pollset work grpc pollset grpc pollset worker long proc self cwd src core lib iomgr pollset cc liblibgrpc ubase uc so run poller void grpc error proc self cwd src core ext filters client channel backup poller cc liblibgrpc uclient uchannel so exec ctx run grpc closure grpc error proc self cwd src core lib iomgr exec ctx cc liblibgrpc ubase uc so grpc core execctx flush proc self cwd src core lib iomgr exec ctx cc liblibgrpc ubase uc so run some timers proc self cwd src core lib iomgr timer manager cc liblibgrpc ubase uc so timer main loop proc self cwd src core lib iomgr timer manager cc liblibgrpc ubase uc so timer thread void proc self cwd src core lib iomgr timer manager cc liblibgrpc ubase uc so grpc core anonymous namespace threadinternalsposix threadinternalsposix char const void void void bool grpc core thread options const lambda void operator void const proc self cwd src core lib gprpp thd posix cc liblibgpr ubase so grpc core anonymous namespace threadinternalsposix threadinternalsposix char const void void void bool grpc core thread options const lambda void invoke void proc self cwd src core lib gprpp thd posix cc liblibgpr ubase so location is heap block of size at allocated by main thread calloc tmp clang build src compiler rt lib tsan rtl tsan interceptors cpp grpclb test gpr zalloc proc self cwd src core lib gpr alloc cc liblibgpr ubase so grpc channel stack builder finish grpc channel stack builder unsigned long int void void grpc error void void proc self cwd src core lib channel channel stack builder cc liblibgrpc ubase uc so grpc channel create with builder grpc channel stack builder grpc channel stack type proc self cwd src core lib surface channel cc liblibgrpc ubase uc so grpc channel create char const grpc channel args const grpc channel stack type grpc transport grpc resource user proc self cwd src core lib surface channel cc liblibgrpc ubase uc so grpc core anonymous namespace createchannel char const grpc channel args const proc self cwd src core ext transport client secure secure channel create cc liblibgrpc utransport uclient usecure so grpc secure channel create proc self cwd src core ext transport client secure secure channel create cc liblibgrpc utransport uclient usecure so grpc impl securechannelcredentials createchannelwithinterceptors std basic string std allocator const grpc impl channelarguments const std vector std allocator proc self cwd src cpp client secure credentials cc liblibgrpc so grpc impl securechannelcredentials createchannelimpl std basic string std allocator const grpc impl channelarguments const proc self cwd src cpp client secure credentials cc liblibgrpc so grpc impl createcustomchannelimpl std basic string std allocator const std shared ptr const grpc impl channelarguments const proc self cwd src cpp client create channel cc liblibgrpc ubase so grpc createcustomchannel std basic string std allocator const std shared ptr const grpc impl channelarguments const proc self cwd include grpcpp create channel h grpclb test grpc testing anonymous namespace resetstub int std basic string std allocator const proc self cwd test cpp grpclb test cc grpclb test grpc testing anonymous namespace setup proc self cwd test cpp grpclb test cc grpclb test void testing internal handlesehexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so void testing internal handleexceptionsinmethodifsupported testing test void testing test char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing test run proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing testinfo run proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing testsuite run proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing internal unittestimpl runalltests proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so bool testing internal handlesehexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so bool testing internal handleexceptionsinmethodifsupported testing internal unittestimpl bool testing internal unittestimpl char const proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so testing unittest run proc self cwd external com github google googletest googletest src gtest cc libexternal scom ugithub ugoogle ugoogletest slibgtest so run all tests proc self cwd external com github google googletest googletest include gtest gtest h grpclb test main proc self cwd test cpp grpclb test cc grpclb test thread grpc global tim tid running created by thread at pthread create tmp clang build src compiler rt lib tsan rtl tsan interceptors cpp grpclb test grpc core anonymous namespace threadinternalsposix threadinternalsposix char const void void void bool grpc core thread options const proc self cwd src core lib gprpp thd posix cc liblibgpr ubase so grpc core thread thread char const void void void bool grpc core thread options const proc self cwd src core lib gprpp thd posix cc liblibgpr ubase so start timer thread and unlock proc self cwd src core lib iomgr timer manager cc liblibgrpc ubase uc so run some timers proc self cwd src core lib iomgr timer manager cc liblibgrpc ubase uc so timer main loop proc self cwd src core lib iomgr timer manager cc liblibgrpc ubase uc so timer thread void proc self cwd src core lib iomgr timer manager cc liblibgrpc ubase uc so grpc core anonymous namespace threadinternalsposix threadinternalsposix char const void void void bool grpc core thread options const lambda void operator void const proc self cwd src core lib gprpp thd posix cc liblibgpr ubase so grpc core anonymous namespace threadinternalsposix threadinternalsposix char const void void void bool grpc core thread options const lambda void invoke void proc self cwd src core lib gprpp thd posix cc liblibgpr ubase so summary threadsanitizer data race usr lib gcc linux gnu include c bits stl tree h in std rb tree std less std allocator size const seen on
0
278,297
30,702,256,935
IssuesEvent
2023-07-27 01:15:19
Nivaskumark/CVE-2020-0114-frameworks_base_afterfix
https://api.github.com/repos/Nivaskumark/CVE-2020-0114-frameworks_base_afterfix
opened
CVE-2023-21117 (High) detected in baseandroid-11.0.0_r39
Mend: dependency security vulnerability
## CVE-2023-21117 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-11.0.0_r39</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/java/com/android/server/am/ActivityManagerService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In registerReceiverWithFeature of ActivityManagerService.java, there is a possible way for isolated processes to register a broadcast receiver due to a permissions bypass. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-263358101 <p>Publish Date: 2023-05-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-21117>CVE-2023-21117</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/frameworks/base/+/8542596db53b1acfb0bf461c93900ff78b34edad">https://android.googlesource.com/platform/frameworks/base/+/8542596db53b1acfb0bf461c93900ff78b34edad</a></p> <p>Release Date: 2023-05-15</p> <p>Fix Resolution: android-13.0.0_r49</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-21117 (High) detected in baseandroid-11.0.0_r39 - ## CVE-2023-21117 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-11.0.0_r39</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/java/com/android/server/am/ActivityManagerService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In registerReceiverWithFeature of ActivityManagerService.java, there is a possible way for isolated processes to register a broadcast receiver due to a permissions bypass. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-263358101 <p>Publish Date: 2023-05-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-21117>CVE-2023-21117</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/frameworks/base/+/8542596db53b1acfb0bf461c93900ff78b34edad">https://android.googlesource.com/platform/frameworks/base/+/8542596db53b1acfb0bf461c93900ff78b34edad</a></p> <p>Release Date: 2023-05-15</p> <p>Fix Resolution: android-13.0.0_r49</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in baseandroid cve high severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in base branch master vulnerable source files services core java com android server am activitymanagerservice java vulnerability details in registerreceiverwithfeature of activitymanagerservice java there is a possible way for isolated processes to register a broadcast receiver due to a permissions bypass this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
0
311,869
26,822,545,382
IssuesEvent
2023-02-02 10:29:02
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] PrioritizedThrottledTaskRunnerTests testFailsTasksOnRejectionOrShutdown failing
:Distributed/Snapshot/Restore >test-failure Team:Distributed
maybe related to https://github.com/elastic/elasticsearch/pull/92627 **Build scan:** https://gradle-enterprise.elastic.co/s/qzk6e7s5deafi/tests/:server:test/org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests/testFailsTasksOnRejectionOrShutdown **Reproduction line:** ``` ./gradlew ':server:test' --tests "org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests.testFailsTasksOnRejectionOrShutdown" -Dtests.seed=A03E55D52C2D5443 -Dtests.locale=es-CL -Dtests.timezone=Africa/Maseru -Druntime.java=17 ``` **Applicable branches:** main **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests&tests.test=testFailsTasksOnRejectionOrShutdown **Failure excerpt:** ``` java.lang.AssertionError: (No message provided) at __randomizedtesting.SeedInfo.seed([A03E55D52C2D5443:1B8402A9B1EC909D]:0) at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests.testFailsTasksOnRejectionOrShutdown(PrioritizedThrottledTaskRunnerTests.java:220) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
1.0
[CI] PrioritizedThrottledTaskRunnerTests testFailsTasksOnRejectionOrShutdown failing - maybe related to https://github.com/elastic/elasticsearch/pull/92627 **Build scan:** https://gradle-enterprise.elastic.co/s/qzk6e7s5deafi/tests/:server:test/org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests/testFailsTasksOnRejectionOrShutdown **Reproduction line:** ``` ./gradlew ':server:test' --tests "org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests.testFailsTasksOnRejectionOrShutdown" -Dtests.seed=A03E55D52C2D5443 -Dtests.locale=es-CL -Dtests.timezone=Africa/Maseru -Druntime.java=17 ``` **Applicable branches:** main **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests&tests.test=testFailsTasksOnRejectionOrShutdown **Failure excerpt:** ``` java.lang.AssertionError: (No message provided) at __randomizedtesting.SeedInfo.seed([A03E55D52C2D5443:1B8402A9B1EC909D]:0) at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at org.elasticsearch.common.util.concurrent.PrioritizedThrottledTaskRunnerTests.testFailsTasksOnRejectionOrShutdown(PrioritizedThrottledTaskRunnerTests.java:220) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
non_process
prioritizedthrottledtaskrunnertests testfailstasksonrejectionorshutdown failing maybe related to build scan reproduction line gradlew server test tests org elasticsearch common util concurrent prioritizedthrottledtaskrunnertests testfailstasksonrejectionorshutdown dtests seed dtests locale es cl dtests timezone africa maseru druntime java applicable branches main reproduces locally no failure history failure excerpt java lang assertionerror no message provided at randomizedtesting seedinfo seed at org junit assert fail assert java at org junit assert asserttrue assert java at org junit assert asserttrue assert java at org elasticsearch common util concurrent prioritizedthrottledtaskrunnertests testfailstasksonrejectionorshutdown prioritizedthrottledtaskrunnertests java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java
0
18,406
24,543,585,365
IssuesEvent
2022-10-12 06:57:46
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Batch processing: Clip raster by mask layer (GDAL)
Processing Bug
### What is the bug or the crash? QGIS adds the Y resolution with a - that makes the batch file not to run because there is no option -2.0. See the code shows `-tr 2.0 -2.0` when I think it should be `-tr 2.0 2.0`. `GDAL command: gdalwarp -overwrite -s_srs EPSG:7856 -t_srs EPSG:7856 -of -tr 2.0 -2.0 -tap -cutline F:/path_to_raster/Clean_raster.shp -cl Clean_raster -crop_to_cutline F:/path_to_raster/Differences_wd.asc F:\path_to_raster\Differences_wd GDAL command output: Process returned error code 1 ERROR 6: Unknown option name '-2.0' Usage: gdalwarp [--help-general] [--formats] [-s_srs srs_def] [-t_srs srs_def] [-to "NAME=VALUE"]* [-vshift | -novshift] [[-s_coord_epoch epoch] | [-t_coord_epoch epoch]] [-order n | -tps | -rpc | -geoloc] [-et err_threshold] [-refine_gcps tolerance [minimum_gcps]] [-te xmin ymin xmax ymax] [-tr xres yres] [-tap] [-ts width height] [-ovr level|AUTO|AUTO-n|NONE] [-wo "NAME=VALUE"] [-ot Byte/Int16/...] [-wt Byte/Int16] [-srcnodata "value [value...]"] [-dstnodata "value [value...]"] -dstalpha [-r resampling_method] [-wm memory_in_mb] [-multi] [-q] [-cutline datasource] [-cl layer] [-cwhere expression] [-csql statement] [-cblend dist_in_pixels] [-crop_to_cutline] [-if format]* [-of format] [-co "NAME=VALUE"]* [-overwrite] [-nomd] [-cvmd meta_conflict_value] [-setci] [-oo NAME=VALUE]* [-doo NAME=VALUE]* srcfile* dstfile Available resampling methods: near (default), bilinear, cubic, cubicspline, lanczos, average, rms, mode, max, min, med, Q1, Q3, sum. Algorithm Clip raster by mask layer correctly executed… Execution completed in 0.19 seconds Results: {'OUTPUT': 'F:\\path_to_raster2\\Differences_wd'} Loading resulting layers Batch execution completed in 1.51 seconds` ### Steps to reproduce the issue Use the tool with two file (Raster ascii) and click on run ### Versions QGIS version 3.22.6-Białowieża QGIS code revision b6f33b1a Compiled against Qt 5.15.2 Running against Qt 5.15.3 Python version 3.9.5 Compiled against GDAL/OGR 3.4.2 Running against GDAL/OGR 3.4.3 PROJ version 9.0.0 EPSG Registry database version v10.054 (2022-02-13) GEOS version 3.10.2-CAPI-1.16.0 SQLite version 3.38.1 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins AusMap 1.1 concavehull 2.0 icsm_ntv2_transformer 1.1.1 MapsPrinter 0.9 mmqgis 2021.9.10 profiletool 4.2.2 quick_map_services 0.19.29 tuflow 3.3 valuetool 3.0.15 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
1.0
Batch processing: Clip raster by mask layer (GDAL) - ### What is the bug or the crash? QGIS adds the Y resolution with a - that makes the batch file not to run because there is no option -2.0. See the code shows `-tr 2.0 -2.0` when I think it should be `-tr 2.0 2.0`. `GDAL command: gdalwarp -overwrite -s_srs EPSG:7856 -t_srs EPSG:7856 -of -tr 2.0 -2.0 -tap -cutline F:/path_to_raster/Clean_raster.shp -cl Clean_raster -crop_to_cutline F:/path_to_raster/Differences_wd.asc F:\path_to_raster\Differences_wd GDAL command output: Process returned error code 1 ERROR 6: Unknown option name '-2.0' Usage: gdalwarp [--help-general] [--formats] [-s_srs srs_def] [-t_srs srs_def] [-to "NAME=VALUE"]* [-vshift | -novshift] [[-s_coord_epoch epoch] | [-t_coord_epoch epoch]] [-order n | -tps | -rpc | -geoloc] [-et err_threshold] [-refine_gcps tolerance [minimum_gcps]] [-te xmin ymin xmax ymax] [-tr xres yres] [-tap] [-ts width height] [-ovr level|AUTO|AUTO-n|NONE] [-wo "NAME=VALUE"] [-ot Byte/Int16/...] [-wt Byte/Int16] [-srcnodata "value [value...]"] [-dstnodata "value [value...]"] -dstalpha [-r resampling_method] [-wm memory_in_mb] [-multi] [-q] [-cutline datasource] [-cl layer] [-cwhere expression] [-csql statement] [-cblend dist_in_pixels] [-crop_to_cutline] [-if format]* [-of format] [-co "NAME=VALUE"]* [-overwrite] [-nomd] [-cvmd meta_conflict_value] [-setci] [-oo NAME=VALUE]* [-doo NAME=VALUE]* srcfile* dstfile Available resampling methods: near (default), bilinear, cubic, cubicspline, lanczos, average, rms, mode, max, min, med, Q1, Q3, sum. Algorithm Clip raster by mask layer correctly executed… Execution completed in 0.19 seconds Results: {'OUTPUT': 'F:\\path_to_raster2\\Differences_wd'} Loading resulting layers Batch execution completed in 1.51 seconds` ### Steps to reproduce the issue Use the tool with two file (Raster ascii) and click on run ### Versions QGIS version 3.22.6-Białowieża QGIS code revision b6f33b1a Compiled against Qt 5.15.2 Running against Qt 5.15.3 Python version 3.9.5 Compiled against GDAL/OGR 3.4.2 Running against GDAL/OGR 3.4.3 PROJ version 9.0.0 EPSG Registry database version v10.054 (2022-02-13) GEOS version 3.10.2-CAPI-1.16.0 SQLite version 3.38.1 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins AusMap 1.1 concavehull 2.0 icsm_ntv2_transformer 1.1.1 MapsPrinter 0.9 mmqgis 2021.9.10 profiletool 4.2.2 quick_map_services 0.19.29 tuflow 3.3 valuetool 3.0.15 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
process
batch processing clip raster by mask layer gdal what is the bug or the crash qgis adds the y resolution with a that makes the batch file not to run because there is no option see the code shows tr when i think it should be tr gdal command gdalwarp overwrite s srs epsg t srs epsg of tr tap cutline f path to raster clean raster shp cl clean raster crop to cutline f path to raster differences wd asc f path to raster differences wd gdal command output process returned error code error unknown option name usage gdalwarp dstalpha srcfile dstfile available resampling methods near default bilinear cubic cubicspline lanczos average rms mode max min med sum algorithm clip raster by mask layer correctly executed… execution completed in seconds results output f path to differences wd loading resulting layers batch execution completed in seconds steps to reproduce the issue use the tool with two file raster ascii and click on run versions qgis version białowieża qgis code revision compiled against qt running against qt python version compiled against gdal ogr running against gdal ogr proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins ausmap concavehull icsm transformer mapsprinter mmqgis profiletool quick map services tuflow valuetool db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
96,423
3,968,288,006
IssuesEvent
2016-05-03 19:07:17
PolarisSS13/Polaris
https://api.github.com/repos/PolarisSS13/Polaris
closed
Some new clothing lacks item sprites
Bug Priority: Low
#### Brief description of the issue Jon did a dumb and the new jeans clothing doesn't have an actual item sprite, but it does have an on-mob sprite. #### What you expected to happen Has sprite. #### What actually happened Orange ? no sprite placeholder. #### Steps to reproduce Put on some grey jeans shorts, for example. #### Additional info: - **Server Revision**: Too lazy to fill out gonna be honest - **Anything else you may wish to add** (Location if it's a mapping issue, etc) I will fix this when I have a few minutes. This is basically to remind me that it needs to be done.
1.0
Some new clothing lacks item sprites - #### Brief description of the issue Jon did a dumb and the new jeans clothing doesn't have an actual item sprite, but it does have an on-mob sprite. #### What you expected to happen Has sprite. #### What actually happened Orange ? no sprite placeholder. #### Steps to reproduce Put on some grey jeans shorts, for example. #### Additional info: - **Server Revision**: Too lazy to fill out gonna be honest - **Anything else you may wish to add** (Location if it's a mapping issue, etc) I will fix this when I have a few minutes. This is basically to remind me that it needs to be done.
non_process
some new clothing lacks item sprites brief description of the issue jon did a dumb and the new jeans clothing doesn t have an actual item sprite but it does have an on mob sprite what you expected to happen has sprite what actually happened orange no sprite placeholder steps to reproduce put on some grey jeans shorts for example additional info server revision too lazy to fill out gonna be honest anything else you may wish to add location if it s a mapping issue etc i will fix this when i have a few minutes this is basically to remind me that it needs to be done
0
35,391
6,459,801,472
IssuesEvent
2017-08-16 01:09:24
DavidAylaian/Carbon
https://api.github.com/repos/DavidAylaian/Carbon
closed
Update README.md to include recent structural changes
priority: low type: documentation
The new [arch](../tree/master/arch) directory is not yet documented in [README.md](../blob/master/README.md).
1.0
Update README.md to include recent structural changes - The new [arch](../tree/master/arch) directory is not yet documented in [README.md](../blob/master/README.md).
non_process
update readme md to include recent structural changes the new tree master arch directory is not yet documented in blob master readme md
0
666,060
22,341,061,742
IssuesEvent
2022-06-15 00:52:03
jordan-sullivan/travel-tracker
https://api.github.com/repos/jordan-sullivan/travel-tracker
closed
It. 2 DASHBBOARD
High Priority UX/UI DOM manipulation MVP
Traveler Interaction As a traveler: I should be able to make a trip request: I will select a date, duration, number of travelers and choose from a list of destinations After making these selections, I should see an estimated cost (with a 10% travel agent fee) for the trip. Once I submit the trip request, it will show on my dashboard as “pending” so that the travel agency can approve or deny it.
1.0
It. 2 DASHBBOARD - Traveler Interaction As a traveler: I should be able to make a trip request: I will select a date, duration, number of travelers and choose from a list of destinations After making these selections, I should see an estimated cost (with a 10% travel agent fee) for the trip. Once I submit the trip request, it will show on my dashboard as “pending” so that the travel agency can approve or deny it.
non_process
it dashbboard traveler interaction as a traveler i should be able to make a trip request i will select a date duration number of travelers and choose from a list of destinations after making these selections i should see an estimated cost with a travel agent fee for the trip once i submit the trip request it will show on my dashboard as “pending” so that the travel agency can approve or deny it
0
586,365
17,576,118,612
IssuesEvent
2021-08-15 16:39:05
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
opened
Satyrs and "true" Demonhood
suggestion :question: lore :books: priority low :grey_exclamation:
<!-- **DO NOT REMOVE PRE-EXISTING LINES** ------------------------------------------------------------------------------------------------------------ --> **Describe your suggestion in full detail below:** Are Satyr's true demons? Becouse i feel like they are more of a Fel Orc kind of being. and yeah i would argue that Xavius is a true demon and/or void being. But i don't think Satyrs pop up anywhere else but on Azeroth (yes i know that outland have some, but those traveled from Azeroth right?) like they even have the theme of the Satyr curse, that is closer to orc's theme rather then ascending to demonhood theme. There are plenty of stuff that point of them being demonic, but that argument can be said for like fel orcs, Felblood elf or Felborne to. An Argument against them being demons are that other Demons are said to be reborn in the Twisting Nether and return upon death.and while Satyrs are described as immortal demons and are classified as such in WoW, none except a few Xavious&Terestian Illhoof (but those have special circumstances) of the killed named satyr have returned yet. It could be thus speculated that satyrs are not true demons or that they are just an exception to the rebirth. So Should we classify them as Demons or should we would classify them the same as we would fel orcs, Felblood elf & Felborne? What do you people think?
1.0
Satyrs and "true" Demonhood - <!-- **DO NOT REMOVE PRE-EXISTING LINES** ------------------------------------------------------------------------------------------------------------ --> **Describe your suggestion in full detail below:** Are Satyr's true demons? Becouse i feel like they are more of a Fel Orc kind of being. and yeah i would argue that Xavius is a true demon and/or void being. But i don't think Satyrs pop up anywhere else but on Azeroth (yes i know that outland have some, but those traveled from Azeroth right?) like they even have the theme of the Satyr curse, that is closer to orc's theme rather then ascending to demonhood theme. There are plenty of stuff that point of them being demonic, but that argument can be said for like fel orcs, Felblood elf or Felborne to. An Argument against them being demons are that other Demons are said to be reborn in the Twisting Nether and return upon death.and while Satyrs are described as immortal demons and are classified as such in WoW, none except a few Xavious&Terestian Illhoof (but those have special circumstances) of the killed named satyr have returned yet. It could be thus speculated that satyrs are not true demons or that they are just an exception to the rebirth. So Should we classify them as Demons or should we would classify them the same as we would fel orcs, Felblood elf & Felborne? What do you people think?
non_process
satyrs and true demonhood do not remove pre existing lines describe your suggestion in full detail below are satyr s true demons becouse i feel like they are more of a fel orc kind of being and yeah i would argue that xavius is a true demon and or void being but i don t think satyrs pop up anywhere else but on azeroth yes i know that outland have some but those traveled from azeroth right like they even have the theme of the satyr curse that is closer to orc s theme rather then ascending to demonhood theme there are plenty of stuff that point of them being demonic but that argument can be said for like fel orcs felblood elf or felborne to an argument against them being demons are that other demons are said to be reborn in the twisting nether and return upon death and while satyrs are described as immortal demons and are classified as such in wow none except a few xavious terestian illhoof but those have special circumstances of the killed named satyr have returned yet it could be thus speculated that satyrs are not true demons or that they are just an exception to the rebirth so should we classify them as demons or should we would classify them the same as we would fel orcs felblood elf felborne what do you people think
0
211,162
16,431,396,409
IssuesEvent
2021-05-20 02:26:29
Max-Edelson/cse110-sp21-group16
https://api.github.com/repos/Max-Edelson/cse110-sp21-group16
closed
Ambiguity in Monthly_Log.js
bug documentation help wanted question
I am trying to create unit tests for Monthly_Log.js rn and I'm looking at ```this.days = [];``` and I have no idea what ```days``` is holding. Is it simply holding a number of days? Or is it holding a day object? If it is only holding a number, what benefit does that provide? Also, didn't we discuss that the monthly log would hold day objects?
1.0
Ambiguity in Monthly_Log.js - I am trying to create unit tests for Monthly_Log.js rn and I'm looking at ```this.days = [];``` and I have no idea what ```days``` is holding. Is it simply holding a number of days? Or is it holding a day object? If it is only holding a number, what benefit does that provide? Also, didn't we discuss that the monthly log would hold day objects?
non_process
ambiguity in monthly log js i am trying to create unit tests for monthly log js rn and i m looking at this days and i have no idea what days is holding is it simply holding a number of days or is it holding a day object if it is only holding a number what benefit does that provide also didn t we discuss that the monthly log would hold day objects
0
498,763
14,432,280,633
IssuesEvent
2020-12-07 01:26:22
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
closed
MSSQL source connector seems broken
area/integration priority/high type/bug
https://analytics.amplitude.com/daxtarity/chart/46qlv54 Stats show that attempts end in failure most of the time This might not be the case, but seems worth checking.
1.0
MSSQL source connector seems broken - https://analytics.amplitude.com/daxtarity/chart/46qlv54 Stats show that attempts end in failure most of the time This might not be the case, but seems worth checking.
non_process
mssql source connector seems broken stats show that attempts end in failure most of the time this might not be the case but seems worth checking
0