Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
169
2,586,826,095
IssuesEvent
2015-02-17 14:49:49
MozillaFoundation/plan
https://api.github.com/repos/MozillaFoundation/plan
closed
Prototype a better etherpad for MoFos
p2 process
Add support for images, formatting and more. ### RACI * Phase: Prototyping * Owner: @openmatt * Decision Maker: @openmatt * Design Lead: not needed * Development Lead: @jdotpz * Quality Verifier: @davidascher * Issues: https://github.com/MozillaFoundation/mofo-lightsaber/issues MoFos use etherpads every day. Let's try to give them a better one. With support for things like images, formatting, better exporting to HTML and markdown, etc. ### Audience * MoFos + the community and people they work with ### Success * Formatting * Images * New ways to find / track pads * Embedding pads. through an iframe, etc. * Better exporting. So that etherpads can be turned into posts, pages, PDFs and tickets, etc. * Stretch goal: make etherpad better so that people aren't as inclined to use Google Docs. Which are hard to find, require permission-setting, and not as open. ### Vision * Updating to the latest instance and experimenting with the right plug-ins should be enough to test. ### Measurement * Success will be measured by happier MoFos. But also: improved behavior. On how we create and share documents and drafts in progress. * More visual thinking. Better visual style and visual language in our communication. Easier passing from etherpads into more readable / hardened artefacts and documents.
1.0
Prototype a better etherpad for MoFos - Add support for images, formatting and more. ### RACI * Phase: Prototyping * Owner: @openmatt * Decision Maker: @openmatt * Design Lead: not needed * Development Lead: @jdotpz * Quality Verifier: @davidascher * Issues: https://github.com/MozillaFoundation/mofo-lightsaber/issues MoFos use etherpads every day. Let's try to give them a better one. With support for things like images, formatting, better exporting to HTML and markdown, etc. ### Audience * MoFos + the community and people they work with ### Success * Formatting * Images * New ways to find / track pads * Embedding pads. through an iframe, etc. * Better exporting. So that etherpads can be turned into posts, pages, PDFs and tickets, etc. * Stretch goal: make etherpad better so that people aren't as inclined to use Google Docs. Which are hard to find, require permission-setting, and not as open. ### Vision * Updating to the latest instance and experimenting with the right plug-ins should be enough to test. ### Measurement * Success will be measured by happier MoFos. But also: improved behavior. On how we create and share documents and drafts in progress. * More visual thinking. Better visual style and visual language in our communication. Easier passing from etherpads into more readable / hardened artefacts and documents.
process
prototype a better etherpad for mofos add support for images formatting and more raci phase prototyping owner openmatt decision maker openmatt design lead not needed development lead jdotpz quality verifier davidascher issues mofos use etherpads every day let s try to give them a better one with support for things like images formatting better exporting to html and markdown etc audience mofos the community and people they work with success formatting images new ways to find track pads embedding pads through an iframe etc better exporting so that etherpads can be turned into posts pages pdfs and tickets etc stretch goal make etherpad better so that people aren t as inclined to use google docs which are hard to find require permission setting and not as open vision updating to the latest instance and experimenting with the right plug ins should be enough to test measurement success will be measured by happier mofos but also improved behavior on how we create and share documents and drafts in progress more visual thinking better visual style and visual language in our communication easier passing from etherpads into more readable hardened artefacts and documents
1
41,252
16,675,835,268
IssuesEvent
2021-06-07 16:03:54
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Alerting] [o11y] Determine how to represent collected metrics
Feature:Actions Feature:Alerting Feature:Task Manager Team:Alerting Services
Relates to https://github.com/elastic/kibana/issues/98902#issuecomment-840698137 Once we determine the set of metrics that are valuable (see https://github.com/elastic/kibana/issues/100675), we want to take a small amount of time to think through how an end user might leverage these metrics. This will most likely be some sort of dashboard, but it will be helpful to know as many specifics as possible to ensure the data is shaped appropriately. For example, if we envision users will only (at least for the short term) be interested in the top n slowest of something, perhaps we can modify data collection to only collect those particular data points (as our current [monitoring metrics](https://github.com/elastic/kibana/blob/master/x-pack/plugins/task_manager/server/MONITORING.md) collect rolling averages so it seems possible to do a little logic before shipping data to only ship top n slowest, or something along those lines)
1.0
[Alerting] [o11y] Determine how to represent collected metrics - Relates to https://github.com/elastic/kibana/issues/98902#issuecomment-840698137 Once we determine the set of metrics that are valuable (see https://github.com/elastic/kibana/issues/100675), we want to take a small amount of time to think through how an end user might leverage these metrics. This will most likely be some sort of dashboard, but it will be helpful to know as many specifics as possible to ensure the data is shaped appropriately. For example, if we envision users will only (at least for the short term) be interested in the top n slowest of something, perhaps we can modify data collection to only collect those particular data points (as our current [monitoring metrics](https://github.com/elastic/kibana/blob/master/x-pack/plugins/task_manager/server/MONITORING.md) collect rolling averages so it seems possible to do a little logic before shipping data to only ship top n slowest, or something along those lines)
non_process
determine how to represent collected metrics relates to once we determine the set of metrics that are valuable see we want to take a small amount of time to think through how an end user might leverage these metrics this will most likely be some sort of dashboard but it will be helpful to know as many specifics as possible to ensure the data is shaped appropriately for example if we envision users will only at least for the short term be interested in the top n slowest of something perhaps we can modify data collection to only collect those particular data points as our current collect rolling averages so it seems possible to do a little logic before shipping data to only ship top n slowest or something along those lines
0
236,138
18,072,033,597
IssuesEvent
2021-09-21 04:45:45
girlscript/winter-of-contributing
https://api.github.com/repos/girlscript/winter-of-contributing
closed
JS: Execution Context in JavaScript
documentation GWOC21 JavaScript Assigned
<hr> ## Description 📜 I'll be explaining briefly about Execution Context in JavaScript in the documentation. <hr> ## Domain of Contribution 📊 - [x] JavaScript <hr> ## Directory `winter-of-contributing/Web_Development/Javascript/Topics/Execution_Context` <hr> ## Formats - [x] Document <hr> I am a GWOC'21 contributor from Batch 1 of JavaScript. Please assign this issue to me. I assure you that this issue's title will be the same as my PR's title as well as I will be Contributing under the Guidelines and Code of Conduct.
1.0
JS: Execution Context in JavaScript - <hr> ## Description 📜 I'll be explaining briefly about Execution Context in JavaScript in the documentation. <hr> ## Domain of Contribution 📊 - [x] JavaScript <hr> ## Directory `winter-of-contributing/Web_Development/Javascript/Topics/Execution_Context` <hr> ## Formats - [x] Document <hr> I am a GWOC'21 contributor from Batch 1 of JavaScript. Please assign this issue to me. I assure you that this issue's title will be the same as my PR's title as well as I will be Contributing under the Guidelines and Code of Conduct.
non_process
js execution context in javascript description 📜 i ll be explaining briefly about execution context in javascript in the documentation domain of contribution 📊 javascript directory winter of contributing web development javascript topics execution context formats document i am a gwoc contributor from batch of javascript please assign this issue to me i assure you that this issue s title will be the same as my pr s title as well as i will be contributing under the guidelines and code of conduct
0
9,173
12,225,909,990
IssuesEvent
2020-05-03 08:13:41
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
missing parent
multi-species process
GO:0140403 should be a child of GO:0052170 (In case it isn't in another ticket, I couldn't see it)
1.0
missing parent - GO:0140403 should be a child of GO:0052170 (In case it isn't in another ticket, I couldn't see it)
process
missing parent go should be a child of go in case it isn t in another ticket i couldn t see it
1
20,732
14,118,320,385
IssuesEvent
2020-11-08 13:10:06
allo-/firefox-profilemaker
https://api.github.com/repos/allo-/firefox-profilemaker
opened
Add "breaks" meta data
Infrastructure
The idea with breaks metadata is to have something like ``` [ { "name": "indexed_db", "type": "boolean", "initial": false, "label": "Disable IndexedDB", "help_text": "(...)", "addons": [], "config": { "dom.indexedDB.enabled": false }, "breaks": { "description": "IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data." "type": "websites" # websites, (firefox) features, etc. "what": [ "All sites, which require indexedDB to work", "Locally saving data on sites, like for example save games or chatlogs", "Example: Twitter now requires IndexedDB to work" ]} } ] ``` Rendered as ``` Disable IndexedDB (description) [ **This setting can break some websites.** IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data. - All sites, which require indexedDB to work - Locally saving data on sites, like for example save games or chatlogs - Example: Twitter now requires IndexedDB to work ] ``` Instead of just having `Disable IndexedDB (breaks things)`. Maybe this could have a short description and a long description and the bullet points in some collapsible element.
1.0
Add "breaks" meta data - The idea with breaks metadata is to have something like ``` [ { "name": "indexed_db", "type": "boolean", "initial": false, "label": "Disable IndexedDB", "help_text": "(...)", "addons": [], "config": { "dom.indexedDB.enabled": false }, "breaks": { "description": "IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data." "type": "websites" # websites, (firefox) features, etc. "what": [ "All sites, which require indexedDB to work", "Locally saving data on sites, like for example save games or chatlogs", "Example: Twitter now requires IndexedDB to work" ]} } ] ``` Rendered as ``` Disable IndexedDB (description) [ **This setting can break some websites.** IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data. - All sites, which require indexedDB to work - Locally saving data on sites, like for example save games or chatlogs - Example: Twitter now requires IndexedDB to work ] ``` Instead of just having `Disable IndexedDB (breaks things)`. Maybe this could have a short description and a long description and the bullet points in some collapsible element.
non_process
add breaks meta data the idea with breaks metadata is to have something like name indexed db type boolean initial false label disable indexeddb help text addons config dom indexeddb enabled false breaks description indexeddb is needed to store indexed data on your pc this could for example be a save game or a chatlog disabling it will not only prevent websites from storing unique identifiers super cookies but also from storing useful data type websites websites firefox features etc what all sites which require indexeddb to work locally saving data on sites like for example save games or chatlogs example twitter now requires indexeddb to work rendered as disable indexeddb description this setting can break some websites indexeddb is needed to store indexed data on your pc this could for example be a save game or a chatlog disabling it will not only prevent websites from storing unique identifiers super cookies but also from storing useful data all sites which require indexeddb to work locally saving data on sites like for example save games or chatlogs example twitter now requires indexeddb to work instead of just having disable indexeddb breaks things maybe this could have a short description and a long description and the bullet points in some collapsible element
0
4,580
7,208,714,287
IssuesEvent
2018-02-07 04:53:06
rust-lang-nursery/futures-rs
https://api.github.com/repos/rust-lang-nursery/futures-rs
closed
Add `Task::park_timeout`
0.1-incompatible 0.2-blocker C-feature-request
I've been rewriting my spmc-broadcast library [`bus`](https://github.com/jonhoo/bus) to expose a `Futures`-based interface rather than `recv/try_recv` (see the [`futures` branch](https://github.com/jonhoo/bus/tree/futures)), and am running into an issue. Due to the way the library works internally, there exists a condition under which the writer won't know that it needs to `unpark` the readers when it does a write. If this happens on the write that fills the buffer, the writer is not blocked on writing further, and the readers aren't aware that they need to start reading. In the non-futures version, this is "fixed" by parking the readers using `thread::park_timeout`. It's not elegant, but it avoids the otherwise costly solution necessary to fix the underlying issue. Unfortunately, with `Futures`, only `task::park` is available, leading to a deadlock in the library whenever the scenario above occurs. Would adding `task::park_timeout` be possible in `Futures`? Are there any major drawbacks to doing so?
True
Add `Task::park_timeout` - I've been rewriting my spmc-broadcast library [`bus`](https://github.com/jonhoo/bus) to expose a `Futures`-based interface rather than `recv/try_recv` (see the [`futures` branch](https://github.com/jonhoo/bus/tree/futures)), and am running into an issue. Due to the way the library works internally, there exists a condition under which the writer won't know that it needs to `unpark` the readers when it does a write. If this happens on the write that fills the buffer, the writer is not blocked on writing further, and the readers aren't aware that they need to start reading. In the non-futures version, this is "fixed" by parking the readers using `thread::park_timeout`. It's not elegant, but it avoids the otherwise costly solution necessary to fix the underlying issue. Unfortunately, with `Futures`, only `task::park` is available, leading to a deadlock in the library whenever the scenario above occurs. Would adding `task::park_timeout` be possible in `Futures`? Are there any major drawbacks to doing so?
non_process
add task park timeout i ve been rewriting my spmc broadcast library to expose a futures based interface rather than recv try recv see the and am running into an issue due to the way the library works internally there exists a condition under which the writer won t know that it needs to unpark the readers when it does a write if this happens on the write that fills the buffer the writer is not blocked on writing further and the readers aren t aware that they need to start reading in the non futures version this is fixed by parking the readers using thread park timeout it s not elegant but it avoids the otherwise costly solution necessary to fix the underlying issue unfortunately with futures only task park is available leading to a deadlock in the library whenever the scenario above occurs would adding task park timeout be possible in futures are there any major drawbacks to doing so
0
585,449
17,497,706,134
IssuesEvent
2021-08-10 04:24:04
code-ready/crc
https://api.github.com/repos/code-ready/crc
closed
[API] Refactor the config to use HTTP verbs
priority/major kind/task
Currently the config endpoint uses `config/get`, `config/set` and `config/unset` to expose the functionality to deal with configuration. However, this seems to be a direct copy of the binary protocol and does not follow HTTP verbs as expected
1.0
[API] Refactor the config to use HTTP verbs - Currently the config endpoint uses `config/get`, `config/set` and `config/unset` to expose the functionality to deal with configuration. However, this seems to be a direct copy of the binary protocol and does not follow HTTP verbs as expected
non_process
refactor the config to use http verbs currently the config endpoint uses config get config set and config unset to expose the functionality to deal with configuration however this seems to be a direct copy of the binary protocol and does not follow http verbs as expected
0
12,177
14,741,968,154
IssuesEvent
2021-01-07 11:28:18
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Dynamic Draft Invoice - Late Fees
anc-process anp-1.5 ant-enhancement
In GitLab by @kdjstudios on Feb 28, 2019, 10:24 **Submitted by:** Kyle **Helpdesk:** NA **Server:** All **Client/Site:** ALL **Account:** ALL **Issue:** IN #1302 we found the possibility that when adjusting items on the draft invoice, the late fee is not updating dynamically. Below is the test case document which contains those scenarios. `Existing Functionality of Late Fee`: [Existing Functionality of Late Fee](https://docs.google.com/spreadsheets/d/1Rx-6jyVIIUqEFDMxnP4Zwe5CIL8Vvcsz-XKGtshFbvg/edit?usp=sharing) `Draft Invoice Test Cases`: [Draft Invoice Test Cases](https://docs.google.com/spreadsheets/d/13pJPni78oybFgp2wN69gSSj7eFGerI6b6m50p8mwn7Q/edit?usp=sharing)
1.0
Dynamic Draft Invoice - Late Fees - In GitLab by @kdjstudios on Feb 28, 2019, 10:24 **Submitted by:** Kyle **Helpdesk:** NA **Server:** All **Client/Site:** ALL **Account:** ALL **Issue:** IN #1302 we found the possibility that when adjusting items on the draft invoice, the late fee is not updating dynamically. Below is the test case document which contains those scenarios. `Existing Functionality of Late Fee`: [Existing Functionality of Late Fee](https://docs.google.com/spreadsheets/d/1Rx-6jyVIIUqEFDMxnP4Zwe5CIL8Vvcsz-XKGtshFbvg/edit?usp=sharing) `Draft Invoice Test Cases`: [Draft Invoice Test Cases](https://docs.google.com/spreadsheets/d/13pJPni78oybFgp2wN69gSSj7eFGerI6b6m50p8mwn7Q/edit?usp=sharing)
process
dynamic draft invoice late fees in gitlab by kdjstudios on feb submitted by kyle helpdesk na server all client site all account all issue in we found the possibility that when adjusting items on the draft invoice the late fee is not updating dynamically below is the test case document which contains those scenarios existing functionality of late fee draft invoice test cases
1
13,161
15,590,007,544
IssuesEvent
2021-03-18 08:49:40
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
prisma migrate gives: Error querying the database: db error: ERROR: syntax error at or near "NOT"
bug/1-repro-available kind/bug process/candidate team/migrations topic: migrate topic: postgresql
## Bug description `prisma migrate dev --preview-feature` is failing with _Error querying the database: db error: ERROR: syntax error at or near "NOT"_ The full debug output is: ``` yarn prisma migrate dev --preview-feature yarn run v1.22.10 warning ../package.json: No license field $ /home/ec2-user/environment/node_modules/.bin/prisma migrate dev --preview-feature prisma:loadEnv project root found at /home/ec2-user/environment/package.json +0ms prisma:tryLoadEnv Environment variables loaded from /home/ec2-user/environment/.env +0ms Environment variables loaded from .env Prisma schema loaded from prisma/schema.prisma Datasource "db": PostgreSQL database "ebdb", schema "public" at "localhost:5432" prisma:migrateEngine:rpc starting migration engine with binary: /home/ec2-user/environment/node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x +0ms prisma:migrateEngine:rpc SENDING RPC CALL {"id":1,"jsonrpc":"2.0","method":"devDiagnostic","params":{"migrationsDirectoryPath":"/home/ec2-user/environment/prisma/migrations"}} +5ms prisma:migrateEngine:stderr Mar 09 12:00:17.017 INFO migration_engine: Starting migration engine RPC server git_hash="da6fafb57b24e0b61ca20960c64e2d41f9e8cff1" +0ms prisma:migrateEngine:stderr Mar 09 12:00:17.031 INFO quaint::single: Starting a postgresql connection. +13ms prisma:migrateEngine:stderr Mar 09 12:00:17.273 INFO DevDiagnostic:calculate_drift:sql_schema_from_migration_history: quaint::single: Starting a postgresql connection. +242ms prisma:migrateEngine:rpc { jsonrpc: '2.0', prisma:migrateEngine:rpc error: prisma:migrateEngine:rpc { code: 4466, prisma:migrateEngine:rpc message: 'An error happened. Check the data field for details.', prisma:migrateEngine:rpc data: prisma:migrateEngine:rpc { is_panic: false, prisma:migrateEngine:rpc message: prisma:migrateEngine:rpc 'Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT"\n 0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history\n at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270\n 1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift\n at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40\n 2: migration_core::api::DevDiagnostic\n at migration-engine/core/src/api.rs:106', prisma:migrateEngine:rpc backtrace: null } }, prisma:migrateEngine:rpc id: 1 } +370ms Error: Error: Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT" 0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270 1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40 2: migration_core::api::DevDiagnostic at migration-engine/core/src/api.rs:106 at Object.registerCallback (/home/ec2-user/environment/node_modules/prisma/build/index.js:55009:26) at MigrateEngine.handleResponse (/home/ec2-user/environment/node_modules/prisma/build/index.js:54882:38) at LineStream.byline_12.default.on (/home/ec2-user/environment/node_modules/prisma/build/index.js:54967:18) at LineStream.emit (events.js:198:13) at LineStream.EventEmitter.emit (domain.js:448:20) at addChunk (_stream_readable.js:288:12) at readableAddChunk (_stream_readable.js:269:11) at LineStream.Readable.push (_stream_readable.js:224:10) at LineStream.Transform.push (_stream_transform.js:151:32) at LineStream._pushBuffer (/home/ec2-user/environment/node_modules/prisma/build/index.js:54719:19) error Command failed with exit code 1. ``` ## How to reproduce I am not entirely sure what the exact steps that lead to the problem are, but hopefully the following is helpful: I had deleted the previous migrations folder by accident, which had a single migration file in it. This migration file set up a basic User model. After seeing the above errors, I did try a complete refresh of the database (dropped and recreated), as well as uninstall/reinstall and reset of prisma. Unfortunately the problem persisted. ## Expected behavior `prisma migrate` to run normally and create the migration sql. ## Prisma information My current prisma schema is: ``` datasource db { provider = "postgresql" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" binaryTargets = ["native"] } model Place { id Int @id @default(autoincrement()) createdAt DateTime @default(now()) updatedAt DateTime @updatedAt place_id String @unique users UserPlace[] } model User { id Int @id @default(autoincrement()) createdAt DateTime @default(now()) updatedAt DateTime @updatedAt email String @unique firstName String lastName String promotion Boolean @default(false) places UserPlace[] } model UserPlace { id Int @id @default(autoincrement()) createdAt DateTime @default(now()) updatedAt DateTime @updatedAt placeId Int? place Place? @relation(fields: [placeId], references: [id]) userId Int? user User? @relation(fields: [userId], references: [id]) } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: Amazon Linux 2 - Database: PostgreSQL (localhost) - Node.js version: v10.23.0 - Prisma version ``` prisma : 2.18.0 @prisma/client : 2.18.0 Current platform : rhel-openssl-1.1.x Query Engine : query-engine da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/query-engine-rhel-openssl-1.1.x) Migration Engine : migration-engine-cli da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x) Introspection Engine : introspection-core da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/introspection-engine-rhel-openssl-1.1.x) Format Binary : prisma-fmt da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/prisma-fmt-rhel-openssl-1.1.x) Studio : 0.356.0 ``` Thanks for your time!
1.0
prisma migrate gives: Error querying the database: db error: ERROR: syntax error at or near "NOT" - ## Bug description `prisma migrate dev --preview-feature` is failing with _Error querying the database: db error: ERROR: syntax error at or near "NOT"_ The full debug output is: ``` yarn prisma migrate dev --preview-feature yarn run v1.22.10 warning ../package.json: No license field $ /home/ec2-user/environment/node_modules/.bin/prisma migrate dev --preview-feature prisma:loadEnv project root found at /home/ec2-user/environment/package.json +0ms prisma:tryLoadEnv Environment variables loaded from /home/ec2-user/environment/.env +0ms Environment variables loaded from .env Prisma schema loaded from prisma/schema.prisma Datasource "db": PostgreSQL database "ebdb", schema "public" at "localhost:5432" prisma:migrateEngine:rpc starting migration engine with binary: /home/ec2-user/environment/node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x +0ms prisma:migrateEngine:rpc SENDING RPC CALL {"id":1,"jsonrpc":"2.0","method":"devDiagnostic","params":{"migrationsDirectoryPath":"/home/ec2-user/environment/prisma/migrations"}} +5ms prisma:migrateEngine:stderr Mar 09 12:00:17.017 INFO migration_engine: Starting migration engine RPC server git_hash="da6fafb57b24e0b61ca20960c64e2d41f9e8cff1" +0ms prisma:migrateEngine:stderr Mar 09 12:00:17.031 INFO quaint::single: Starting a postgresql connection. +13ms prisma:migrateEngine:stderr Mar 09 12:00:17.273 INFO DevDiagnostic:calculate_drift:sql_schema_from_migration_history: quaint::single: Starting a postgresql connection. +242ms prisma:migrateEngine:rpc { jsonrpc: '2.0', prisma:migrateEngine:rpc error: prisma:migrateEngine:rpc { code: 4466, prisma:migrateEngine:rpc message: 'An error happened. Check the data field for details.', prisma:migrateEngine:rpc data: prisma:migrateEngine:rpc { is_panic: false, prisma:migrateEngine:rpc message: prisma:migrateEngine:rpc 'Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT"\n 0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history\n at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270\n 1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift\n at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40\n 2: migration_core::api::DevDiagnostic\n at migration-engine/core/src/api.rs:106', prisma:migrateEngine:rpc backtrace: null } }, prisma:migrateEngine:rpc id: 1 } +370ms Error: Error: Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT" 0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270 1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40 2: migration_core::api::DevDiagnostic at migration-engine/core/src/api.rs:106 at Object.registerCallback (/home/ec2-user/environment/node_modules/prisma/build/index.js:55009:26) at MigrateEngine.handleResponse (/home/ec2-user/environment/node_modules/prisma/build/index.js:54882:38) at LineStream.byline_12.default.on (/home/ec2-user/environment/node_modules/prisma/build/index.js:54967:18) at LineStream.emit (events.js:198:13) at LineStream.EventEmitter.emit (domain.js:448:20) at addChunk (_stream_readable.js:288:12) at readableAddChunk (_stream_readable.js:269:11) at LineStream.Readable.push (_stream_readable.js:224:10) at LineStream.Transform.push (_stream_transform.js:151:32) at LineStream._pushBuffer (/home/ec2-user/environment/node_modules/prisma/build/index.js:54719:19) error Command failed with exit code 1. ``` ## How to reproduce I am not entirely sure what the exact steps that lead to the problem are, but hopefully the following is helpful: I had deleted the previous migrations folder by accident, which had a single migration file in it. This migration file set up a basic User model. After seeing the above errors, I did try a complete refresh of the database (dropped and recreated), as well as uninstall/reinstall and reset of prisma. Unfortunately the problem persisted. ## Expected behavior `prisma migrate` to run normally and create the migration sql. ## Prisma information My current prisma schema is: ``` datasource db { provider = "postgresql" url = env("DATABASE_URL") } generator client { provider = "prisma-client-js" binaryTargets = ["native"] } model Place { id Int @id @default(autoincrement()) createdAt DateTime @default(now()) updatedAt DateTime @updatedAt place_id String @unique users UserPlace[] } model User { id Int @id @default(autoincrement()) createdAt DateTime @default(now()) updatedAt DateTime @updatedAt email String @unique firstName String lastName String promotion Boolean @default(false) places UserPlace[] } model UserPlace { id Int @id @default(autoincrement()) createdAt DateTime @default(now()) updatedAt DateTime @updatedAt placeId Int? place Place? @relation(fields: [placeId], references: [id]) userId Int? user User? @relation(fields: [userId], references: [id]) } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: Amazon Linux 2 - Database: PostgreSQL (localhost) - Node.js version: v10.23.0 - Prisma version ``` prisma : 2.18.0 @prisma/client : 2.18.0 Current platform : rhel-openssl-1.1.x Query Engine : query-engine da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/query-engine-rhel-openssl-1.1.x) Migration Engine : migration-engine-cli da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x) Introspection Engine : introspection-core da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/introspection-engine-rhel-openssl-1.1.x) Format Binary : prisma-fmt da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/prisma-fmt-rhel-openssl-1.1.x) Studio : 0.356.0 ``` Thanks for your time!
process
prisma migrate gives error querying the database db error error syntax error at or near not bug description prisma migrate dev preview feature is failing with error querying the database db error error syntax error at or near not the full debug output is yarn prisma migrate dev preview feature yarn run warning package json no license field home user environment node modules bin prisma migrate dev preview feature prisma loadenv project root found at home user environment package json prisma tryloadenv environment variables loaded from home user environment env environment variables loaded from env prisma schema loaded from prisma schema prisma datasource db postgresql database ebdb schema public at localhost prisma migrateengine rpc starting migration engine with binary home user environment node modules prisma engines migration engine rhel openssl x prisma migrateengine rpc sending rpc call id jsonrpc method devdiagnostic params migrationsdirectorypath home user environment prisma migrations prisma migrateengine stderr mar info migration engine starting migration engine rpc server git hash prisma migrateengine stderr mar info quaint single starting a postgresql connection prisma migrateengine stderr mar info devdiagnostic calculate drift sql schema from migration history quaint single starting a postgresql connection prisma migrateengine rpc jsonrpc prisma migrateengine rpc error prisma migrateengine rpc code prisma migrateengine rpc message an error happened check the data field for details prisma migrateengine rpc data prisma migrateengine rpc is panic false prisma migrateengine rpc message prisma migrateengine rpc database error error querying the database db error error syntax error at or near not n sql migration connector flavour postgres sql schema from migration history n at migration engine connectors sql migration connector src flavour postgres rs n sql migration connector sql database migration inferrer calculate drift n at migration engine connectors sql migration connector src sql database migration inferrer rs n migration core api devdiagnostic n at migration engine core src api rs prisma migrateengine rpc backtrace null prisma migrateengine rpc id error error database error error querying the database db error error syntax error at or near not sql migration connector flavour postgres sql schema from migration history at migration engine connectors sql migration connector src flavour postgres rs sql migration connector sql database migration inferrer calculate drift at migration engine connectors sql migration connector src sql database migration inferrer rs migration core api devdiagnostic at migration engine core src api rs at object registercallback home user environment node modules prisma build index js at migrateengine handleresponse home user environment node modules prisma build index js at linestream byline default on home user environment node modules prisma build index js at linestream emit events js at linestream eventemitter emit domain js at addchunk stream readable js at readableaddchunk stream readable js at linestream readable push stream readable js at linestream transform push stream transform js at linestream pushbuffer home user environment node modules prisma build index js error command failed with exit code how to reproduce i am not entirely sure what the exact steps that lead to the problem are but hopefully the following is helpful i had deleted the previous migrations folder by accident which had a single migration file in it this migration file set up a basic user model after seeing the above errors i did try a complete refresh of the database dropped and recreated as well as uninstall reinstall and reset of prisma unfortunately the problem persisted expected behavior prisma migrate to run normally and create the migration sql prisma information my current prisma schema is datasource db provider postgresql url env database url generator client provider prisma client js binarytargets model place id int id default autoincrement createdat datetime default now updatedat datetime updatedat place id string unique users userplace model user id int id default autoincrement createdat datetime default now updatedat datetime updatedat email string unique firstname string lastname string promotion boolean default false places userplace model userplace id int id default autoincrement createdat datetime default now updatedat datetime updatedat placeid int place place relation fields references userid int user user relation fields references environment setup os amazon linux database postgresql localhost node js version prisma version prisma prisma client current platform rhel openssl x query engine query engine at node modules prisma engines query engine rhel openssl x migration engine migration engine cli at node modules prisma engines migration engine rhel openssl x introspection engine introspection core at node modules prisma engines introspection engine rhel openssl x format binary prisma fmt at node modules prisma engines prisma fmt rhel openssl x studio thanks for your time
1
20,051
26,540,017,103
IssuesEvent
2023-01-19 18:27:29
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
opened
Add ability for plug-in packages to register custom data structures (entities)
type - enhancement stage - planning level - difficult f - processing f - API f - plugins feature type - internal maintenance
Related to-do items: - [ ] include schema during file writes; required to allow reference counting - [ ] automatic version compatibility with schemas; know whether a schema is compatible with an older version - [ ] fully qualified type names; instead of context sensitive entity types which are difficult to implement - [ ] reference count non-root items (i.e. an interval shared by EELS and a graphic)
1.0
Add ability for plug-in packages to register custom data structures (entities) - Related to-do items: - [ ] include schema during file writes; required to allow reference counting - [ ] automatic version compatibility with schemas; know whether a schema is compatible with an older version - [ ] fully qualified type names; instead of context sensitive entity types which are difficult to implement - [ ] reference count non-root items (i.e. an interval shared by EELS and a graphic)
process
add ability for plug in packages to register custom data structures entities related to do items include schema during file writes required to allow reference counting automatic version compatibility with schemas know whether a schema is compatible with an older version fully qualified type names instead of context sensitive entity types which are difficult to implement reference count non root items i e an interval shared by eels and a graphic
1
385,398
11,420,271,319
IssuesEvent
2020-02-03 09:47:27
nhn/tui.grid
https://api.github.com/repos/nhn/tui.grid
closed
vue-grid 4.x 에서 서버 연동
4.x Enhancement Priority: Normal Question
<!-- To make it easier for us to help you, please include as much useful information as possible. Useful Links: - tutorial: https://github.com/nhn/tui.grid/tree/master/docs - API/Example: https://nhn.github.io/tui.grid/latest Before opening a new issue, please search existing issues https://github.com/nhn/tui.grid/issues --> **Summary** 4.x에서 서버연동을 해보고 있습니다. 아래는 현재 서버 연동 시 리턴 규격입니다. [https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md](https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md) 기준 ``` { "result": true, "data": { "contents": [], "pagination": { "page": 1, "totalCount": 100 } } } ``` 공통 규격이 있는 상태에서 그리드 용도의 api만 규격에 맞게 하기는 비효율인 것 같아서요, 위 규격이 아닌 커스텀한 데이터를 할당할 수 있는지요? ex) 아래 규격 중 data 를 할당 ``` { code: 200, message: "정상 처리 되었습니다", data: { result: true, data: { contents: ......, pagination: { ....... } } } } ``` **Screenshots** If applicable, add screenshots to help explain your question. **Version** toast-ui.vue-grid 2.0.1 **Additional context** 추가로 문의드립니다. 1. 서버사이드 구성 시 data값을 아래와 같이 했더니 데이터는 잘 나옵니다만 Invalid prop: type check failed for prop "data". Expected Array, got Object 라는 오류가 발생합니다. 어떻게 할당을 해야 하는지요? 2. 페이지 구성이 되지 않고 있습니다. 어디가 잘못된 건지 알려주시면 감사하겠습니다 ㅠ.ㅠ ``` <template> <grid ref="tuiGrid" :data="gridProps.data" :columns="gridProps.columns" :options="gridProps.options"/> </template> <script> import 'tui-grid/dist/tui-grid.css' import { Grid } from '@toast-ui/vue-grid' export default { components: { 'grid': Grid }, created() { this.gridProps = { data: { api: { readData: { url: '/api/grid/test', method: 'GET' }, } }, pageOptions: { perPage: 5 }, pagination: true, columns: [ { name: "name", header: "이름", align: "center", sortable: true }, { name: "age", header: "나이", align: "center", sortable: true }, { name: "addr", header: "주소", align: "center" } ] } }, methods: { } } </script> <style> </style> ```
1.0
vue-grid 4.x 에서 서버 연동 - <!-- To make it easier for us to help you, please include as much useful information as possible. Useful Links: - tutorial: https://github.com/nhn/tui.grid/tree/master/docs - API/Example: https://nhn.github.io/tui.grid/latest Before opening a new issue, please search existing issues https://github.com/nhn/tui.grid/issues --> **Summary** 4.x에서 서버연동을 해보고 있습니다. 아래는 현재 서버 연동 시 리턴 규격입니다. [https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md](https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md) 기준 ``` { "result": true, "data": { "contents": [], "pagination": { "page": 1, "totalCount": 100 } } } ``` 공통 규격이 있는 상태에서 그리드 용도의 api만 규격에 맞게 하기는 비효율인 것 같아서요, 위 규격이 아닌 커스텀한 데이터를 할당할 수 있는지요? ex) 아래 규격 중 data 를 할당 ``` { code: 200, message: "정상 처리 되었습니다", data: { result: true, data: { contents: ......, pagination: { ....... } } } } ``` **Screenshots** If applicable, add screenshots to help explain your question. **Version** toast-ui.vue-grid 2.0.1 **Additional context** 추가로 문의드립니다. 1. 서버사이드 구성 시 data값을 아래와 같이 했더니 데이터는 잘 나옵니다만 Invalid prop: type check failed for prop "data". Expected Array, got Object 라는 오류가 발생합니다. 어떻게 할당을 해야 하는지요? 2. 페이지 구성이 되지 않고 있습니다. 어디가 잘못된 건지 알려주시면 감사하겠습니다 ㅠ.ㅠ ``` <template> <grid ref="tuiGrid" :data="gridProps.data" :columns="gridProps.columns" :options="gridProps.options"/> </template> <script> import 'tui-grid/dist/tui-grid.css' import { Grid } from '@toast-ui/vue-grid' export default { components: { 'grid': Grid }, created() { this.gridProps = { data: { api: { readData: { url: '/api/grid/test', method: 'GET' }, } }, pageOptions: { perPage: 5 }, pagination: true, columns: [ { name: "name", header: "이름", align: "center", sortable: true }, { name: "age", header: "나이", align: "center", sortable: true }, { name: "addr", header: "주소", align: "center" } ] } }, methods: { } } </script> <style> </style> ```
non_process
vue grid x 에서 서버 연동 to make it easier for us to help you please include as much useful information as possible useful links tutorial api example before opening a new issue please search existing issues summary x에서 서버연동을 해보고 있습니다 아래는 현재 서버 연동 시 리턴 규격입니다 기준 result true data contents pagination page totalcount 공통 규격이 있는 상태에서 그리드 용도의 api만 규격에 맞게 하기는 비효율인 것 같아서요 위 규격이 아닌 커스텀한 데이터를 할당할 수 있는지요 ex 아래 규격 중 data 를 할당 code message 정상 처리 되었습니다 data result true data contents pagination screenshots if applicable add screenshots to help explain your question version toast ui vue grid additional context 추가로 문의드립니다 서버사이드 구성 시 data값을 아래와 같이 했더니 데이터는 잘 나옵니다만 invalid prop type check failed for prop data expected array got object 라는 오류가 발생합니다 어떻게 할당을 해야 하는지요 페이지 구성이 되지 않고 있습니다 어디가 잘못된 건지 알려주시면 감사하겠습니다 ㅠ ㅠ import tui grid dist tui grid css import grid from toast ui vue grid export default components grid grid created this gridprops data api readdata url api grid test method get pageoptions perpage pagination true columns name name header 이름 align center sortable true name age header 나이 align center sortable true name addr header 주소 align center methods
0
630,658
20,115,963,097
IssuesEvent
2022-02-07 19:32:51
IDAES/examples-pse
https://api.github.com/repos/IDAES/examples-pse
closed
More failures in example notebooks
bug Priority:High
The recent merge of https://github.com/IDAES/idaes-pse/pull/627 has revealed additional issues with the HDS flowsheet example (Examples/Flowsheets/HDA_flowsheet_with_distillation_solution_testing.ipynb). It appears the reactor initialization step is/was failing and that it had not been noticed up until now - now that the CSTR initialization routine raises an `InitializationError` if the final solve fails to converge this notebook is now failing. We need to look at this model once again and determine what the issue is and try to improve the models performance (hopefully through scaling). As an absolute last resort, we can wrap the CSTR initialization step in a `try/except` statement to skip the error, but that should be avoided as this is our flagship tutorial example.
1.0
More failures in example notebooks - The recent merge of https://github.com/IDAES/idaes-pse/pull/627 has revealed additional issues with the HDS flowsheet example (Examples/Flowsheets/HDA_flowsheet_with_distillation_solution_testing.ipynb). It appears the reactor initialization step is/was failing and that it had not been noticed up until now - now that the CSTR initialization routine raises an `InitializationError` if the final solve fails to converge this notebook is now failing. We need to look at this model once again and determine what the issue is and try to improve the models performance (hopefully through scaling). As an absolute last resort, we can wrap the CSTR initialization step in a `try/except` statement to skip the error, but that should be avoided as this is our flagship tutorial example.
non_process
more failures in example notebooks the recent merge of has revealed additional issues with the hds flowsheet example examples flowsheets hda flowsheet with distillation solution testing ipynb it appears the reactor initialization step is was failing and that it had not been noticed up until now now that the cstr initialization routine raises an initializationerror if the final solve fails to converge this notebook is now failing we need to look at this model once again and determine what the issue is and try to improve the models performance hopefully through scaling as an absolute last resort we can wrap the cstr initialization step in a try except statement to skip the error but that should be avoided as this is our flagship tutorial example
0
460,029
13,203,457,661
IssuesEvent
2020-08-14 14:12:24
novelis-prod/Digital-CoE-Operations-Data---Public
https://api.github.com/repos/novelis-prod/Digital-CoE-Operations-Data---Public
closed
Pinda tables not partitioned or contain more than just partitions within opsentprodg2
Plant: Pinda Priority #1
Went to look at which dates the pinda tables had within opsentprodg2 and found that many tables are no longer partitioned by anything, or contained partitions and parquet files outside of those partitions ![image](https://user-images.githubusercontent.com/58708978/88690538-c4624580-d0c9-11ea-9123-9794528ccd63.png) ![image](https://user-images.githubusercontent.com/58708978/88690581-cd531700-d0c9-11ea-8b78-3214dd93a4a2.png) ![image](https://user-images.githubusercontent.com/58708978/88690616-d6dc7f00-d0c9-11ea-9f6d-88180f169774.png) ![image](https://user-images.githubusercontent.com/58708978/88690641-de9c2380-d0c9-11ea-9df9-123f2c945b0a.png) ![image](https://user-images.githubusercontent.com/58708978/88690682-e78cf500-d0c9-11ea-9448-3de8c44b4ea3.png) ![image](https://user-images.githubusercontent.com/58708978/88690727-f1aef380-d0c9-11ea-8b1c-454029682fae.png) ![image](https://user-images.githubusercontent.com/58708978/88690744-fa072e80-d0c9-11ea-9a2f-002a04886ffd.png) ![image](https://user-images.githubusercontent.com/58708978/88690776-02f80000-d0ca-11ea-8dee-71a7e15835af.png) ![image](https://user-images.githubusercontent.com/58708978/88690819-0f7c5880-d0ca-11ea-9e55-aaac13d23bc2.png) ![image](https://user-images.githubusercontent.com/58708978/88690848-16a36680-d0ca-11ea-93a4-30d40dfbfa26.png)
1.0
Pinda tables not partitioned or contain more than just partitions within opsentprodg2 - Went to look at which dates the pinda tables had within opsentprodg2 and found that many tables are no longer partitioned by anything, or contained partitions and parquet files outside of those partitions ![image](https://user-images.githubusercontent.com/58708978/88690538-c4624580-d0c9-11ea-9123-9794528ccd63.png) ![image](https://user-images.githubusercontent.com/58708978/88690581-cd531700-d0c9-11ea-8b78-3214dd93a4a2.png) ![image](https://user-images.githubusercontent.com/58708978/88690616-d6dc7f00-d0c9-11ea-9f6d-88180f169774.png) ![image](https://user-images.githubusercontent.com/58708978/88690641-de9c2380-d0c9-11ea-9df9-123f2c945b0a.png) ![image](https://user-images.githubusercontent.com/58708978/88690682-e78cf500-d0c9-11ea-9448-3de8c44b4ea3.png) ![image](https://user-images.githubusercontent.com/58708978/88690727-f1aef380-d0c9-11ea-8b1c-454029682fae.png) ![image](https://user-images.githubusercontent.com/58708978/88690744-fa072e80-d0c9-11ea-9a2f-002a04886ffd.png) ![image](https://user-images.githubusercontent.com/58708978/88690776-02f80000-d0ca-11ea-8dee-71a7e15835af.png) ![image](https://user-images.githubusercontent.com/58708978/88690819-0f7c5880-d0ca-11ea-9e55-aaac13d23bc2.png) ![image](https://user-images.githubusercontent.com/58708978/88690848-16a36680-d0ca-11ea-93a4-30d40dfbfa26.png)
non_process
pinda tables not partitioned or contain more than just partitions within went to look at which dates the pinda tables had within and found that many tables are no longer partitioned by anything or contained partitions and parquet files outside of those partitions
0
189,129
22,046,987,266
IssuesEvent
2022-05-30 03:39:49
madhans23/linux-4.1.15
https://api.github.com/repos/madhans23/linux-4.1.15
closed
CVE-2016-6213 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed
security vulnerability
## CVE-2016-6213 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fs/namespace.c in the Linux kernel before 4.9 does not restrict how many mounts may exist in a mount namespace, which allows local users to cause a denial of service (memory consumption and deadlock) via MS_BIND mount system calls, as demonstrated by a loop that triggers exponential growth in the number of mounts. <p>Publish Date: 2016-12-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6213>CVE-2016-6213</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6213">https://nvd.nist.gov/vuln/detail/CVE-2016-6213</a></p> <p>Release Date: 2016-12-28</p> <p>Fix Resolution: 4.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-6213 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2016-6213 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fs/namespace.c in the Linux kernel before 4.9 does not restrict how many mounts may exist in a mount namespace, which allows local users to cause a denial of service (memory consumption and deadlock) via MS_BIND mount system calls, as demonstrated by a loop that triggers exponential growth in the number of mounts. <p>Publish Date: 2016-12-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6213>CVE-2016-6213</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6213">https://nvd.nist.gov/vuln/detail/CVE-2016-6213</a></p> <p>Release Date: 2016-12-28</p> <p>Fix Resolution: 4.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files fs pnode c vulnerability details fs namespace c in the linux kernel before does not restrict how many mounts may exist in a mount namespace which allows local users to cause a denial of service memory consumption and deadlock via ms bind mount system calls as demonstrated by a loop that triggers exponential growth in the number of mounts publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
175,760
6,553,706,095
IssuesEvent
2017-09-06 00:23:53
opencurrents/opencurrents
https://api.github.com/repos/opencurrents/opencurrents
closed
upcoming-events: Events only displaying for admins of the org who created event
priority high
Volunteers are not seeing any events
1.0
upcoming-events: Events only displaying for admins of the org who created event - Volunteers are not seeing any events
non_process
upcoming events events only displaying for admins of the org who created event volunteers are not seeing any events
0
18,660
24,581,464,637
IssuesEvent
2022-10-13 15:55:36
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[FHIR] Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file
Bug P0 Response datastore Process: Fixed Process: Tested dev
AR: Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file ER: Questionnaires resources > JSON > 'Title' key should get displayed in the JSON file
2.0
[FHIR] Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file - AR: Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file ER: Questionnaires resources > JSON > 'Title' key should get displayed in the JSON file
process
questionnaires resources json title key is not getting displayed in the json file ar questionnaires resources json title key is not getting displayed in the json file er questionnaires resources json title key should get displayed in the json file
1
114,119
14,531,533,997
IssuesEvent
2020-12-14 20:56:29
retaildevcrews/ngsa
https://api.github.com/repos/retaildevcrews/ngsa
closed
Review Data Separation Design
DesignReview PlatValidApp ValidDemo
## Description What: - Review and discuss design considerations from the data separation spike #168 Why: - Drive understanding of new design - Collect any feedback prior to incorporating into codebase When: - ASAP Where: - ngsa app (C#, TS?) ## Tasks - [x] Discuss and Review design #197 - [ ] Document design #196 ## Acceptance Criteria - Meeting had and information delivered - Any critical questions addressed - Design documented - ## Constraints - N/A ## References: - References #8
1.0
Review Data Separation Design - ## Description What: - Review and discuss design considerations from the data separation spike #168 Why: - Drive understanding of new design - Collect any feedback prior to incorporating into codebase When: - ASAP Where: - ngsa app (C#, TS?) ## Tasks - [x] Discuss and Review design #197 - [ ] Document design #196 ## Acceptance Criteria - Meeting had and information delivered - Any critical questions addressed - Design documented - ## Constraints - N/A ## References: - References #8
non_process
review data separation design description what review and discuss design considerations from the data separation spike why drive understanding of new design collect any feedback prior to incorporating into codebase when asap where ngsa app c ts tasks discuss and review design document design acceptance criteria meeting had and information delivered any critical questions addressed design documented constraints n a references references
0
446,472
31,478,604,463
IssuesEvent
2023-08-30 12:30:26
TycheSoftwares/woocommerce-abandoned-cart
https://api.github.com/repos/TycheSoftwares/woocommerce-abandoned-cart
closed
Update the screenshots in the documentation.
type: update documentation
**Describe the bug** Update the screenshots in the documentation as per the current UI of tabs. https://www.tychesoftwares.com/docs/docs/abandoned-cart-for-woocommerce-lite/capturing-abandoned-carts/
1.0
Update the screenshots in the documentation. - **Describe the bug** Update the screenshots in the documentation as per the current UI of tabs. https://www.tychesoftwares.com/docs/docs/abandoned-cart-for-woocommerce-lite/capturing-abandoned-carts/
non_process
update the screenshots in the documentation describe the bug update the screenshots in the documentation as per the current ui of tabs
0
16,858
22,139,784,934
IssuesEvent
2022-06-03 05:06:53
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Remove reflection from general and Azure Service Bus-specific message routing
area:message-processing
### Discussed in https://github.com/arcus-azure/arcus.messaging/discussions/261 <div type='discussions-op-text'> <sup>Originally posted by **fgheysels** March 24, 2022</sup> Not an idea for a new feature, but an idea for a possible improvement. As mentionned in this PR #260 , the code to retrieve registered `IMessageHandler` implementations is not that stable. We should find a way to improve this. @stijnmoreels's comment copied from the mentionned PR: > You'll probably wonder why we have this reflection code in the MessageHandler file. This is bc we were faced with a problem when implementing the message routing. The IServceProvider doesn't support open generic interfaces and since our IMessageHandler<,> is such an open interface, there is no build-in way to extract all service registrations with a generic message type. This message handler class will do this for us. It searches for the 'engine' and 'call site factory' on the service provider and extracts all the message handler interfaces. Later on, we can determine if there exists and message handler that can process the incoming message. > >We can off course discuss this later, or come up with something else. For now, this was the best we could do. Unfortunately. I was wondering if we couldn't work around this without using reflection. Maybe it is an idea to introduce a non-generic interface (empty) `IMessageHandler`. The `IMessageHandler<,>` would then derive from that non-generic `IMessageHandler` interface, like this: ```csharp public interface IMessageHandler { } public interface IMessageHandler<in TMessage, in TMessageContext> : IMessageHandler where TMessageContext : MessageContext { ... } ``` The code for retrieving registered message-handlers could then work on retrieving types that are assignable from the base-interface `IMessageHandler`. Once we have retrieved those types, then we need of course an additional filter: we only want to take into consideration those types that are also assignable to the IMessageHandler interface which declares type parameters.</div>
1.0
Remove reflection from general and Azure Service Bus-specific message routing - ### Discussed in https://github.com/arcus-azure/arcus.messaging/discussions/261 <div type='discussions-op-text'> <sup>Originally posted by **fgheysels** March 24, 2022</sup> Not an idea for a new feature, but an idea for a possible improvement. As mentionned in this PR #260 , the code to retrieve registered `IMessageHandler` implementations is not that stable. We should find a way to improve this. @stijnmoreels's comment copied from the mentionned PR: > You'll probably wonder why we have this reflection code in the MessageHandler file. This is bc we were faced with a problem when implementing the message routing. The IServceProvider doesn't support open generic interfaces and since our IMessageHandler<,> is such an open interface, there is no build-in way to extract all service registrations with a generic message type. This message handler class will do this for us. It searches for the 'engine' and 'call site factory' on the service provider and extracts all the message handler interfaces. Later on, we can determine if there exists and message handler that can process the incoming message. > >We can off course discuss this later, or come up with something else. For now, this was the best we could do. Unfortunately. I was wondering if we couldn't work around this without using reflection. Maybe it is an idea to introduce a non-generic interface (empty) `IMessageHandler`. The `IMessageHandler<,>` would then derive from that non-generic `IMessageHandler` interface, like this: ```csharp public interface IMessageHandler { } public interface IMessageHandler<in TMessage, in TMessageContext> : IMessageHandler where TMessageContext : MessageContext { ... } ``` The code for retrieving registered message-handlers could then work on retrieving types that are assignable from the base-interface `IMessageHandler`. Once we have retrieved those types, then we need of course an additional filter: we only want to take into consideration those types that are also assignable to the IMessageHandler interface which declares type parameters.</div>
process
remove reflection from general and azure service bus specific message routing discussed in originally posted by fgheysels march not an idea for a new feature but an idea for a possible improvement as mentionned in this pr the code to retrieve registered imessagehandler implementations is not that stable we should find a way to improve this stijnmoreels s comment copied from the mentionned pr you ll probably wonder why we have this reflection code in the messagehandler file this is bc we were faced with a problem when implementing the message routing the iservceprovider doesn t support open generic interfaces and since our imessagehandler is such an open interface there is no build in way to extract all service registrations with a generic message type this message handler class will do this for us it searches for the engine and call site factory on the service provider and extracts all the message handler interfaces later on we can determine if there exists and message handler that can process the incoming message we can off course discuss this later or come up with something else for now this was the best we could do unfortunately i was wondering if we couldn t work around this without using reflection maybe it is an idea to introduce a non generic interface empty imessagehandler the imessagehandler would then derive from that non generic imessagehandler interface like this csharp public interface imessagehandler public interface imessagehandler imessagehandler where tmessagecontext messagecontext the code for retrieving registered message handlers could then work on retrieving types that are assignable from the base interface imessagehandler once we have retrieved those types then we need of course an additional filter we only want to take into consideration those types that are also assignable to the imessagehandler interface which declares type parameters
1
7,134
10,278,478,759
IssuesEvent
2019-08-25 14:45:57
nextmoov/nextmoov
https://api.github.com/repos/nextmoov/nextmoov
closed
Getting Started Dev Templates
#Dev Tools & Processes
-[] nodejs vanilla esNext -[] nodejs vanilla esNext + express -[] F-E vanilla esNext -[] F-E vanilla esNext + react -[] F-E vanilla esNext + react+ redux
1.0
Getting Started Dev Templates - -[] nodejs vanilla esNext -[] nodejs vanilla esNext + express -[] F-E vanilla esNext -[] F-E vanilla esNext + react -[] F-E vanilla esNext + react+ redux
process
getting started dev templates nodejs vanilla esnext nodejs vanilla esnext express f e vanilla esnext f e vanilla esnext react f e vanilla esnext react redux
1
747,221
26,077,585,092
IssuesEvent
2022-12-24 19:57:20
zigtools/zls
https://api.github.com/repos/zigtools/zls
closed
Another integer overflow in `fullPtrType`
bug priority:high fuzzing result
### Zig Version 0.11.0-dev.782+0b4461d97 ### Zig Language Server Version 0.11.0-dev.72+3526f5f ### Steps to Reproduce Open the following file: ```zig **CapsuleHeader, .description = "SPIR-V version = version. var x69: u1 = undefined; addcarryxU64(&x112, &x113, x110); try std.testing.expectEqualSlices(u8, &bytes, &[_]u8{}, it.peek().?, "d"); length is less than `locked`. /// For example, instead of file action type, but it is safe for user protocol = extern struct { pub const CGT_SYM = "__vdso_clock_getres50(clk_id: c_int) isize; pub extern "c" fn posix_spawnattr_setguardsize(attr: *posix_spawnattr_t = extern struct { // Ensure that there must be freed before the window is not implementation failed with @asyncCall: HANDLE, // TODO: Remove and return error.Permissions for Embedded ProcessForWsWatch(hProcess: [*]u8, .NOMEM => return error.SystemResources, const math = std.meta; const ld_info.name)); } else |err| @as(i32, -1), @as(i32, 4), queue.removeMin(), 1); /// with the state as bytes in POSIX message into memory copy, but you are connectionRefused; } pub fn proc_exit(0); .allocator }; } // Unimplementation information about to do it. .tag = .decl_between_source }; } /// Compute x*y (mod L) pub const p: anyframe }, .TrueLiteral1; alg_version) |ver| { const sqe_recv = net.AddressInUse, var u2a = U2{ .C = 0 }; proc: extern "user32" fn Dispatch" { \\\\ switch (w.args_sizes_get_ret = os.wasi.clock_time_get_ns, end_ix = original_deps[0].name, "A")); try bit_stream.readBits(u5, 2, &out_bits)) { else extern union { return self.unmanaged{ .integer startup key or signal_t = 1; var off_out); .fresh => log.info("FrameworkWeak(self: *Self, elem: T) []T { t4 = t4.mul(Y3); if (len == b.len) : (offset < try seekable, var k: i32 = undefined behavior of non-instructions", try expectFmt("hello world!", "{[1]s} world{[punctuation is correct type. ATOMIC_LOCKS_NOT_STARTED = 0x00000004; /// Get the pointer to these bits < byte_size; } /// Delete a direct writes if more than the anyframe->ReturnType(comptime buffer size (in bytes const s_per_min; pub const ELFDATA2MSB => .Big, .dependencies = featureSet(&[_]Feature{ \\\\ ); } test "y_object_id_t = i32; pub const value: [2]u64, /// therefore the equality here on, we only need to double free &out_port, // `new_len` must be called on this command is of type `T` has a bug", break; } } return self.unmanaged(Node.Index) full.FnProto { var r = try std.testing.expect(mem.eql(u8, it.next()) { \\\\}; const len = Codecs{ .android) struct { : [number] "{rbp}" (@enumToInt(Feature.lvi_load_hardening", field_state = .u8; } /// A system cannot be deleted. RPC_S_OBJECT_PATH_SYNTAX_BAD => unreachable, return error.EndOfStream { inline data. /// When trying next members_range.end], .level_3, return p.addNode(.{ } .tail_call = 0, while (i > 0) { try testing.expectEqualSlices(u8, expected` == hex(`input`) where || is concatenation fn parse(s: []const u8, dir_list: std.c.MACH_TASK_BASIC_INFORMATION = 0xC01E0354, NOT_CONTAINER_STATE_VETOED = 0x8000000000000000000000000000000000000000000000000000000, 0x22, 0x68, 0x65, 0x6e, 0xe8, 0x90, 0x27, 0x3f, 0xdf, 0xfd616b15, 0x2f501ec8, 0xad0552ab, 0x07cfe89cf1bd76ac, \\\\["\\u200B"] ); } test "y_array_expr, r: []Limb) usize { \\\\pub extern "c" fn thread_id: u64, .{ .key = "042927b10512bae3eddcfe467828128bad2903269919f7086069c8c4df6c732838c7787964eaac00e5921fb1498a60f4606766b3d9685001558d1a974e7341513e", .msg = "313233343030", .sig = "3026090180022100b329f479a2bbd0a5c384ee1493b1f5186a87139cac5df4087c134b49156847db", .result = try os.openbsd.FUTEX_WAIT takes no requeue address + end_offset, shdr.sh_size) { try testing.expect(isWhitespace. FWP_DYNAMIC .llvm_name = "arm1020e", .step = Step.init(.run, name, ".debug_info = @typeInfo(T).Int.bits) return Literal e exponent -= exponent_fast_path) catch null; E{ .part = 0xc07, .m32 = &A64.cortex_a8", \\\\ # Return the_copy.dependencies = featureSet(&[_]Feature{ } } .rclass, fp16fml, // added the state byte_count = fields_delimiterArrayListUnmanaged.clearRetainingCapacityContext(key: [32]u8) Curve25519BasePointPc[0..9] else pc: { else if (mem.eql(u8, name, ".")) { .llvm_name = null, single_threaded) { .fast_skip_hashshing = 6, ftruncate(u64, x >> 4)); mem.swap(Self, self.fd, sub_path_w, new_dir_mode); .unsigned, math.log1p(x + x * r32(x * x); if (i >= iovecs[i].iov_len = @sizeOf(c_int) aligned_addr == 0 and i & 2 != 0, a.isSet(0)); // present. MUTUAL_AUTH_FAILED = 1634, log2_buf_align: u8, ret_addr: ?*sockaddr, addr_size: *usize, value: T, endian); another allocator: mem.Allocator); .clwb, try std.testing.allocator, u); self.rd_pos..self.window_end == 0) try p.warnMsg(.{ .@"3e3r2", return badDwarf(); /// Write `string` to `writer` is not valid. INVALID_ICON_HANDLE = IOC_OUT | IOC_WS2 | 27; pub const WM_CAP_FILE_SAVEASW = 0x0479; pub const IOCOUTQ = if (is_array) |_, bit_index + 1; \\\\ asm volatile ("svc #0" .main_token = p.assertTuple(.{u32}, Tuple(&[_]type{ f16, f32, f64, f80, f128, c_long, 124); h.update(options.allocator, 1); .time_low = 0xa7af67cb, handler support mirroring display mode in word width pub const ABRT = 6; /// but this extra buffer len /// dropped a clear the base-2 shift. This swaps the literal = .llu }; } } CPU, .description = "Output struct { return self.endianSwap(); inline while (i < 16) : (j += 1) { } return FileSource, state.end_index -= 1; } error.Utf8OverlongEncodingError.NoSpaceLeft, .avr0, addcarryxU64(&x79, &x80, x1, (x13 & 0xc7634d81f4372ddf); if (index += 1; .pku, 0xb0ba0da5bb600397, sgetmask = 200, .strx = undefined; subborrowxU64(&x57, &x58, x56, x19, x21); } One, error.FileDescriptor could not used. ABORTED = 0xC00000CA, .INVAL => error.AccessDenied, .float1e2, try expectEqualStrings(self.bits.set(index)) != 0; fbs.pos += n; .Shader, SPV_INTEL_unstructure pointer to the specified in this binary is foreign binaries. /// Format = struct { fd: os.fd_t, /// The ArrayList(u8).init(key, npub); }, return Self{ const b10 = (s0 -% y0) >> 63; pub const IP_WFP_REDIRECT_CONTEXT_NOT_FOUND => return result.tag = .caret; for (section) |only_section: std.os.iovec; extern "wasi_snapshot_preview1" fn fd_filestat_set_times(fd: fd_t, buf: [*]u8, fast_gather, .btst16, /// IPsec dropped when MH_PREBOUND_DYLIB = (0x18 | LC_REQ_DYLD), \\\\ "tab": "\\t", .clock_settime64 = 421, PIPE_NOWAIT = 0xC00002B1, f32x4_relaxed_laneselect = 142, secs: u17, // max avail throughput }); '=' => { \\\\ , & && &= ? < <= << .LOOP => return GzipStream(allocator, maxInt(u3)) } .description = "Reserve X17", } const c = p.t.mul(z_inv); // T*z_inv .fullfp16, }, .Dynamic => try p.warn(.expected_continuation); } /// `queryCapsuleCapability StoragePushConstant(in_stream.readIntBig(u128, &l); compiler of Zig, using this function subborrowxU64(&x69, &x70, x1, (arg2[3])); try testing.expectEqual(linux.getuid()); const fd = @intCast(MinInt)` }, tp.* = .{ // Note: can't splice mem.writeIntLittle(u16, utf16le_buf.len, mem.page_size, shstr64.sh_offset).* += base_addr = @ptrCast(*FailingAlloc(alloc, "\\"foo\\\ bar\\"))); prev: ?*LinkMap, xsavec, inherit it /// The checks are a /// primitive integer. WSA_E_CANCELLED = 0xC0220013, /// multiple process argv/env pub const b = Complex = cmath.Complex; /// PrivateUsage: SIZE_T, PUT, \\\\} IV[1], }, if (args.len != 64) return self.resize(self.items[i], items[2]); var kf = invln2 * x; defer tmp_dir.dir.makePath(self.builder.dupe(phrase), } } } fn renderExpressions: ?[]const u8 = null, defer a.deinit(self: FileSource, } }; const index_size = dyn_size = hasher.final(); } fn renderToken(ais, tree, body, space); .vector_enhancement for slice[read_idx] == '.') orelse return error.SystemResources, if (!node.view.pending.fetchSub(5, ordering), const window parameter to IndexedArray(EnumIndexer(Indexer); defer std.testing; const WINAPI or STDCALL. /// {Missing System and return mem.eql(u8, content = true; return @enumToInt(Feature.avx512f, .Keyword_do, Insn.ld_imm(0), slab("24641172998046401", -56), slab("2229658838863212", -132), .NoReturn: void, R9: DWORD64, var x43: u64 = undefined; cmovznzU64(&x216, x215, x212, x235); var e = DeflateFast{ iu += 0x3F800000) { return pcMul16(pc: *const fn (c_int) + 1, .space); defer tmp.cleanup(); : [number], %%eax while (true) { try checkNext(&p, .String = .{ .StringLiteral(c_ulong) c_long, 50); pub const POOL = 49; pub const LUP_RETURN_ALIASES = 1024; pub const bit = @intCast(ShelfIndex) usize { .user_data = user_data: u32, mips5_32r2, v.* = Stack.init(); /// File exists and the password, ring.deinit(); } const ad = "Additional div" { const x54 = (x42 + x53); var buf: [8192]u8 = undefined; var enabled, to run on Valgrind() usize { writercv: pthread_create(name: *const [16 * count]u8 = undefined. Note: microsoft/microsoft.com/en-us/library/windows/ole32.zig"); pub const imported global_instance = @ptrCast(*anyopaque = blk: { const bytes = [_:0]u16{0} ** codegen and codegen_code_count]u8 = undefined; var i: usize) void { } .treap = Treap(u64, u64).init(r, y - y); artifact.target.isDarwin(tag: Tag, arch: Cpu.Arch) Machine Scheduling policy failure = .{ .path = "test_file.txt"); h = Blake2b(Box.nonce_length]u8 = [_]u8{}; defer tmp.cleanup(); const fn (context, lessThan(i32, 0, 5) catch return null; } // FileTime(*FILETIME, lpWSAData: *WSABUF, dwProviderFlags, /// The new Mutable and this map &out_port, \\\\push $0 } .avoid_partial; builder.pathJoin(&lib_list[0] = p[0]; const iovec = std.os.windows) { .next = undefined; cmovznzU64(&x295, &x296, x258, x255); var buf1: [1]u8 = undefined; mulxU64(&x331, &x332, x5, 0xffffffffffffff); var acc = st.atime(), else => {}, }); \\\\} return q; } return readlinkW(file_path); var t2: ?TokenIndex, space: Space = enum(u32) { p.count = 0, destination addcarryxU64(&x47, &x48, x1, (arg2[2]), (arg3[2]), x18); var result: [len]CpuFeature.ccdp)] = .{ .disk_designatorWindows("\\\\\\\\unc\\\\share\\\\foo\\\\", "\\\\\\\\foo\\\\bar", "c:\\\\foo\\\\tmp.3\\\\", "..\\\\baz"); if (name.len); } 0x30, 0x2f83071f53f325dd, f32_store = 0x38, flags: u32) c_long, 76); pub const s1 = try toPosixPath(sub_dir_path); if (callconv(WINAPI) NTSTATUS; pub extern "user32" fn GetCurrentDirectory as the index in `entries` must be used instead"), ); try testing.expect(isSorted(u8, "zyxw", {}, desc_u8)); defer q.deinit(); attr.map_elem.key = @ptrCast(*[digest_length; .AGAIN => unreachable, } return stream.offsetTrue(); : "cc", "memory" ); } test "Reader.skipBytes(1, rem_key[8..]), seed), .Plus => "+", return path; while (lane < threads); } ``` ### Expected Behavior No crash. ### Actual Behavior ```log thread 9704 panic: integer overflow C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:637:50: 0x7ff61da206e9 in firstToken (zls.exe.obj) => switch (token_tags[main_token - 1]) { ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:38:37: 0x7ff61db6b717 in fullPtrType (zls.exe.obj) const end = tree.firstToken(info.child_type); ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:98:27: 0x7ff61daaebc9 in ptrTypeAligned (zls.exe.obj) .child_type = data.rhs, ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:924:44: 0x7ff61da173f7 in ptrType (zls.exe.obj) .ptr_type_aligned => ptrTypeAligned(tree, node), ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3046:59: 0x7ff61d9f83fd in makeScopeInternal (zls.exe.obj) const ptr_type: Ast.full.PtrType = ast.ptrType(tree, node_idx).?; ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff61d9f8166 in makeScopeInternal (zls.exe.obj) try makeScopeInternal(allocator, context, field.ast.type_expr); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff61da96a93 in makeInnerScope (zls.exe.obj) try makeScopeInternal(allocator, context, decl); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff61d9f4df6 in makeScopeInternal (zls.exe.obj) try makeInnerScope(allocator, context, node_idx); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff61d9f41cd in makeDocumentScope (zls.exe.obj) .enums = &document_scope.enum_completions, ^ C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:208:83: 0x7ff61d9f933e in refreshDocument (zls.exe.obj) var new_document_scope = try analysis.makeDocumentScope(self.allocator, handle.tree); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1881:53: 0x7ff61d9fb309 in changeDocumentHandler__anon_12775 (zls.exe.obj) try server.document_store.refreshDocument(handle.uri, new_text); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff61da5677c in processJsonRpc__anon_10459 (zls.exe.obj) method_info[2](server, writer, id, request_obj) catch |err| { ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff61da5d754 in loop (zls.exe.obj) try server.processJsonRpc(writer, buffer); ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff61da5dbd2 in main (zls.exe.obj) try loop(&server); ^ C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff61da5e077 in WinStartup (zls.exe.obj) std.debug.maybeEnableSegfaultHandler(); ^ ???:?:?: 0x7ffab1d9559f in ??? (???) ???:?:?: 0x7ffab2c0485a in ??? (???) ```
1.0
Another integer overflow in `fullPtrType` - ### Zig Version 0.11.0-dev.782+0b4461d97 ### Zig Language Server Version 0.11.0-dev.72+3526f5f ### Steps to Reproduce Open the following file: ```zig **CapsuleHeader, .description = "SPIR-V version = version. var x69: u1 = undefined; addcarryxU64(&x112, &x113, x110); try std.testing.expectEqualSlices(u8, &bytes, &[_]u8{}, it.peek().?, "d"); length is less than `locked`. /// For example, instead of file action type, but it is safe for user protocol = extern struct { pub const CGT_SYM = "__vdso_clock_getres50(clk_id: c_int) isize; pub extern "c" fn posix_spawnattr_setguardsize(attr: *posix_spawnattr_t = extern struct { // Ensure that there must be freed before the window is not implementation failed with @asyncCall: HANDLE, // TODO: Remove and return error.Permissions for Embedded ProcessForWsWatch(hProcess: [*]u8, .NOMEM => return error.SystemResources, const math = std.meta; const ld_info.name)); } else |err| @as(i32, -1), @as(i32, 4), queue.removeMin(), 1); /// with the state as bytes in POSIX message into memory copy, but you are connectionRefused; } pub fn proc_exit(0); .allocator }; } // Unimplementation information about to do it. .tag = .decl_between_source }; } /// Compute x*y (mod L) pub const p: anyframe }, .TrueLiteral1; alg_version) |ver| { const sqe_recv = net.AddressInUse, var u2a = U2{ .C = 0 }; proc: extern "user32" fn Dispatch" { \\\\ switch (w.args_sizes_get_ret = os.wasi.clock_time_get_ns, end_ix = original_deps[0].name, "A")); try bit_stream.readBits(u5, 2, &out_bits)) { else extern union { return self.unmanaged{ .integer startup key or signal_t = 1; var off_out); .fresh => log.info("FrameworkWeak(self: *Self, elem: T) []T { t4 = t4.mul(Y3); if (len == b.len) : (offset < try seekable, var k: i32 = undefined behavior of non-instructions", try expectFmt("hello world!", "{[1]s} world{[punctuation is correct type. ATOMIC_LOCKS_NOT_STARTED = 0x00000004; /// Get the pointer to these bits < byte_size; } /// Delete a direct writes if more than the anyframe->ReturnType(comptime buffer size (in bytes const s_per_min; pub const ELFDATA2MSB => .Big, .dependencies = featureSet(&[_]Feature{ \\\\ ); } test "y_object_id_t = i32; pub const value: [2]u64, /// therefore the equality here on, we only need to double free &out_port, // `new_len` must be called on this command is of type `T` has a bug", break; } } return self.unmanaged(Node.Index) full.FnProto { var r = try std.testing.expect(mem.eql(u8, it.next()) { \\\\}; const len = Codecs{ .android) struct { : [number] "{rbp}" (@enumToInt(Feature.lvi_load_hardening", field_state = .u8; } /// A system cannot be deleted. RPC_S_OBJECT_PATH_SYNTAX_BAD => unreachable, return error.EndOfStream { inline data. /// When trying next members_range.end], .level_3, return p.addNode(.{ } .tail_call = 0, while (i > 0) { try testing.expectEqualSlices(u8, expected` == hex(`input`) where || is concatenation fn parse(s: []const u8, dir_list: std.c.MACH_TASK_BASIC_INFORMATION = 0xC01E0354, NOT_CONTAINER_STATE_VETOED = 0x8000000000000000000000000000000000000000000000000000000, 0x22, 0x68, 0x65, 0x6e, 0xe8, 0x90, 0x27, 0x3f, 0xdf, 0xfd616b15, 0x2f501ec8, 0xad0552ab, 0x07cfe89cf1bd76ac, \\\\["\\u200B"] ); } test "y_array_expr, r: []Limb) usize { \\\\pub extern "c" fn thread_id: u64, .{ .key = "042927b10512bae3eddcfe467828128bad2903269919f7086069c8c4df6c732838c7787964eaac00e5921fb1498a60f4606766b3d9685001558d1a974e7341513e", .msg = "313233343030", .sig = "3026090180022100b329f479a2bbd0a5c384ee1493b1f5186a87139cac5df4087c134b49156847db", .result = try os.openbsd.FUTEX_WAIT takes no requeue address + end_offset, shdr.sh_size) { try testing.expect(isWhitespace. FWP_DYNAMIC .llvm_name = "arm1020e", .step = Step.init(.run, name, ".debug_info = @typeInfo(T).Int.bits) return Literal e exponent -= exponent_fast_path) catch null; E{ .part = 0xc07, .m32 = &A64.cortex_a8", \\\\ # Return the_copy.dependencies = featureSet(&[_]Feature{ } } .rclass, fp16fml, // added the state byte_count = fields_delimiterArrayListUnmanaged.clearRetainingCapacityContext(key: [32]u8) Curve25519BasePointPc[0..9] else pc: { else if (mem.eql(u8, name, ".")) { .llvm_name = null, single_threaded) { .fast_skip_hashshing = 6, ftruncate(u64, x >> 4)); mem.swap(Self, self.fd, sub_path_w, new_dir_mode); .unsigned, math.log1p(x + x * r32(x * x); if (i >= iovecs[i].iov_len = @sizeOf(c_int) aligned_addr == 0 and i & 2 != 0, a.isSet(0)); // present. MUTUAL_AUTH_FAILED = 1634, log2_buf_align: u8, ret_addr: ?*sockaddr, addr_size: *usize, value: T, endian); another allocator: mem.Allocator); .clwb, try std.testing.allocator, u); self.rd_pos..self.window_end == 0) try p.warnMsg(.{ .@"3e3r2", return badDwarf(); /// Write `string` to `writer` is not valid. INVALID_ICON_HANDLE = IOC_OUT | IOC_WS2 | 27; pub const WM_CAP_FILE_SAVEASW = 0x0479; pub const IOCOUTQ = if (is_array) |_, bit_index + 1; \\\\ asm volatile ("svc #0" .main_token = p.assertTuple(.{u32}, Tuple(&[_]type{ f16, f32, f64, f80, f128, c_long, 124); h.update(options.allocator, 1); .time_low = 0xa7af67cb, handler support mirroring display mode in word width pub const ABRT = 6; /// but this extra buffer len /// dropped a clear the base-2 shift. This swaps the literal = .llu }; } } CPU, .description = "Output struct { return self.endianSwap(); inline while (i < 16) : (j += 1) { } return FileSource, state.end_index -= 1; } error.Utf8OverlongEncodingError.NoSpaceLeft, .avr0, addcarryxU64(&x79, &x80, x1, (x13 & 0xc7634d81f4372ddf); if (index += 1; .pku, 0xb0ba0da5bb600397, sgetmask = 200, .strx = undefined; subborrowxU64(&x57, &x58, x56, x19, x21); } One, error.FileDescriptor could not used. ABORTED = 0xC00000CA, .INVAL => error.AccessDenied, .float1e2, try expectEqualStrings(self.bits.set(index)) != 0; fbs.pos += n; .Shader, SPV_INTEL_unstructure pointer to the specified in this binary is foreign binaries. /// Format = struct { fd: os.fd_t, /// The ArrayList(u8).init(key, npub); }, return Self{ const b10 = (s0 -% y0) >> 63; pub const IP_WFP_REDIRECT_CONTEXT_NOT_FOUND => return result.tag = .caret; for (section) |only_section: std.os.iovec; extern "wasi_snapshot_preview1" fn fd_filestat_set_times(fd: fd_t, buf: [*]u8, fast_gather, .btst16, /// IPsec dropped when MH_PREBOUND_DYLIB = (0x18 | LC_REQ_DYLD), \\\\ "tab": "\\t", .clock_settime64 = 421, PIPE_NOWAIT = 0xC00002B1, f32x4_relaxed_laneselect = 142, secs: u17, // max avail throughput }); '=' => { \\\\ , & && &= ? < <= << .LOOP => return GzipStream(allocator, maxInt(u3)) } .description = "Reserve X17", } const c = p.t.mul(z_inv); // T*z_inv .fullfp16, }, .Dynamic => try p.warn(.expected_continuation); } /// `queryCapsuleCapability StoragePushConstant(in_stream.readIntBig(u128, &l); compiler of Zig, using this function subborrowxU64(&x69, &x70, x1, (arg2[3])); try testing.expectEqual(linux.getuid()); const fd = @intCast(MinInt)` }, tp.* = .{ // Note: can't splice mem.writeIntLittle(u16, utf16le_buf.len, mem.page_size, shstr64.sh_offset).* += base_addr = @ptrCast(*FailingAlloc(alloc, "\\"foo\\\ bar\\"))); prev: ?*LinkMap, xsavec, inherit it /// The checks are a /// primitive integer. WSA_E_CANCELLED = 0xC0220013, /// multiple process argv/env pub const b = Complex = cmath.Complex; /// PrivateUsage: SIZE_T, PUT, \\\\} IV[1], }, if (args.len != 64) return self.resize(self.items[i], items[2]); var kf = invln2 * x; defer tmp_dir.dir.makePath(self.builder.dupe(phrase), } } } fn renderExpressions: ?[]const u8 = null, defer a.deinit(self: FileSource, } }; const index_size = dyn_size = hasher.final(); } fn renderToken(ais, tree, body, space); .vector_enhancement for slice[read_idx] == '.') orelse return error.SystemResources, if (!node.view.pending.fetchSub(5, ordering), const window parameter to IndexedArray(EnumIndexer(Indexer); defer std.testing; const WINAPI or STDCALL. /// {Missing System and return mem.eql(u8, content = true; return @enumToInt(Feature.avx512f, .Keyword_do, Insn.ld_imm(0), slab("24641172998046401", -56), slab("2229658838863212", -132), .NoReturn: void, R9: DWORD64, var x43: u64 = undefined; cmovznzU64(&x216, x215, x212, x235); var e = DeflateFast{ iu += 0x3F800000) { return pcMul16(pc: *const fn (c_int) + 1, .space); defer tmp.cleanup(); : [number], %%eax while (true) { try checkNext(&p, .String = .{ .StringLiteral(c_ulong) c_long, 50); pub const POOL = 49; pub const LUP_RETURN_ALIASES = 1024; pub const bit = @intCast(ShelfIndex) usize { .user_data = user_data: u32, mips5_32r2, v.* = Stack.init(); /// File exists and the password, ring.deinit(); } const ad = "Additional div" { const x54 = (x42 + x53); var buf: [8192]u8 = undefined; var enabled, to run on Valgrind() usize { writercv: pthread_create(name: *const [16 * count]u8 = undefined. Note: microsoft/microsoft.com/en-us/library/windows/ole32.zig"); pub const imported global_instance = @ptrCast(*anyopaque = blk: { const bytes = [_:0]u16{0} ** codegen and codegen_code_count]u8 = undefined; var i: usize) void { } .treap = Treap(u64, u64).init(r, y - y); artifact.target.isDarwin(tag: Tag, arch: Cpu.Arch) Machine Scheduling policy failure = .{ .path = "test_file.txt"); h = Blake2b(Box.nonce_length]u8 = [_]u8{}; defer tmp.cleanup(); const fn (context, lessThan(i32, 0, 5) catch return null; } // FileTime(*FILETIME, lpWSAData: *WSABUF, dwProviderFlags, /// The new Mutable and this map &out_port, \\\\push $0 } .avoid_partial; builder.pathJoin(&lib_list[0] = p[0]; const iovec = std.os.windows) { .next = undefined; cmovznzU64(&x295, &x296, x258, x255); var buf1: [1]u8 = undefined; mulxU64(&x331, &x332, x5, 0xffffffffffffff); var acc = st.atime(), else => {}, }); \\\\} return q; } return readlinkW(file_path); var t2: ?TokenIndex, space: Space = enum(u32) { p.count = 0, destination addcarryxU64(&x47, &x48, x1, (arg2[2]), (arg3[2]), x18); var result: [len]CpuFeature.ccdp)] = .{ .disk_designatorWindows("\\\\\\\\unc\\\\share\\\\foo\\\\", "\\\\\\\\foo\\\\bar", "c:\\\\foo\\\\tmp.3\\\\", "..\\\\baz"); if (name.len); } 0x30, 0x2f83071f53f325dd, f32_store = 0x38, flags: u32) c_long, 76); pub const s1 = try toPosixPath(sub_dir_path); if (callconv(WINAPI) NTSTATUS; pub extern "user32" fn GetCurrentDirectory as the index in `entries` must be used instead"), ); try testing.expect(isSorted(u8, "zyxw", {}, desc_u8)); defer q.deinit(); attr.map_elem.key = @ptrCast(*[digest_length; .AGAIN => unreachable, } return stream.offsetTrue(); : "cc", "memory" ); } test "Reader.skipBytes(1, rem_key[8..]), seed), .Plus => "+", return path; while (lane < threads); } ``` ### Expected Behavior No crash. ### Actual Behavior ```log thread 9704 panic: integer overflow C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:637:50: 0x7ff61da206e9 in firstToken (zls.exe.obj) => switch (token_tags[main_token - 1]) { ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:38:37: 0x7ff61db6b717 in fullPtrType (zls.exe.obj) const end = tree.firstToken(info.child_type); ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:98:27: 0x7ff61daaebc9 in ptrTypeAligned (zls.exe.obj) .child_type = data.rhs, ^ C:\Programming\Zig\buzz\repos\zls\src\ast.zig:924:44: 0x7ff61da173f7 in ptrType (zls.exe.obj) .ptr_type_aligned => ptrTypeAligned(tree, node), ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3046:59: 0x7ff61d9f83fd in makeScopeInternal (zls.exe.obj) const ptr_type: Ast.full.PtrType = ast.ptrType(tree, node_idx).?; ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff61d9f8166 in makeScopeInternal (zls.exe.obj) try makeScopeInternal(allocator, context, field.ast.type_expr); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff61da96a93 in makeInnerScope (zls.exe.obj) try makeScopeInternal(allocator, context, decl); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff61d9f4df6 in makeScopeInternal (zls.exe.obj) try makeInnerScope(allocator, context, node_idx); ^ C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff61d9f41cd in makeDocumentScope (zls.exe.obj) .enums = &document_scope.enum_completions, ^ C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:208:83: 0x7ff61d9f933e in refreshDocument (zls.exe.obj) var new_document_scope = try analysis.makeDocumentScope(self.allocator, handle.tree); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1881:53: 0x7ff61d9fb309 in changeDocumentHandler__anon_12775 (zls.exe.obj) try server.document_store.refreshDocument(handle.uri, new_text); ^ C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff61da5677c in processJsonRpc__anon_10459 (zls.exe.obj) method_info[2](server, writer, id, request_obj) catch |err| { ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff61da5d754 in loop (zls.exe.obj) try server.processJsonRpc(writer, buffer); ^ C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff61da5dbd2 in main (zls.exe.obj) try loop(&server); ^ C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff61da5e077 in WinStartup (zls.exe.obj) std.debug.maybeEnableSegfaultHandler(); ^ ???:?:?: 0x7ffab1d9559f in ??? (???) ???:?:?: 0x7ffab2c0485a in ??? (???) ```
non_process
another integer overflow in fullptrtype zig version dev zig language server version dev steps to reproduce open the following file zig capsuleheader description spir v version version var undefined try std testing expectequalslices bytes it peek d length is less than locked for example instead of file action type but it is safe for user protocol extern struct pub const cgt sym vdso clock clk id c int isize pub extern c fn posix spawnattr setguardsize attr posix spawnattr t extern struct ensure that there must be freed before the window is not implementation failed with asynccall handle todo remove and return error permissions for embedded processforwswatch hprocess nomem return error systemresources const math std meta const ld info name else err as as queue removemin with the state as bytes in posix message into memory copy but you are connectionrefused pub fn proc exit allocator unimplementation information about to do it tag decl between source compute x y mod l pub const p anyframe alg version ver const sqe recv net addressinuse var c proc extern fn dispatch switch w args sizes get ret os wasi clock time get ns end ix original deps name a try bit stream readbits out bits else extern union return self unmanaged integer startup key or signal t var off out fresh log info frameworkweak self self elem t t mul if len b len offset try seekable var k undefined behavior of non instructions try expectfmt hello world s world punctuation is correct type atomic locks not started get the pointer to these bits byte size delete a direct writes if more than the anyframe returntype comptime buffer size in bytes const s per min pub const big dependencies featureset feature test y object id t pub const value therefore the equality here on we only need to double free out port new len must be called on this command is of type t has a bug break return self unmanaged node index full fnproto var r try std testing expect mem eql it next const len codecs android struct rbp enumtoint feature lvi load hardening field state a system cannot be deleted rpc s object path syntax bad unreachable return error endofstream inline data when trying next members range end level return p addnode tail call while i try testing expectequalslices expected hex input where is concatenation fn parse s const dir list std c mach task basic information not container state vetoed test y array expr r limb usize pub extern c fn thread id key msg sig result try os openbsd futex wait takes no requeue address end offset shdr sh size try testing expect iswhitespace fwp dynamic llvm name step step init run name debug info typeinfo t int bits return literal e exponent exponent fast path catch null e part cortex return the copy dependencies featureset feature rclass added the state byte count fields delimiterarraylistunmanaged clearretainingcapacitycontext key else pc else if mem eql name llvm name null single threaded fast skip hashshing ftruncate x mem swap self self fd sub path w new dir mode unsigned math x x x x if i iovecs iov len sizeof c int aligned addr and i a isset present mutual auth failed buf align ret addr sockaddr addr size usize value t endian another allocator mem allocator clwb try std testing allocator u self rd pos self window end try p warnmsg return baddwarf write string to writer is not valid invalid icon handle ioc out ioc pub const wm cap file saveasw pub const iocoutq if is array bit index asm volatile svc main token p asserttuple tuple type c long h update options allocator time low handler support mirroring display mode in word width pub const abrt but this extra buffer len dropped a clear the base shift this swaps the literal llu cpu description output struct return self endianswap inline while i j return filesource state end index error nospaceleft if index pku sgetmask strx undefined one error filedescriptor could not used aborted inval error accessdenied try expectequalstrings self bits set index fbs pos n shader spv intel unstructure pointer to the specified in this binary is foreign binaries format struct fd os fd t the arraylist init key npub return self const pub const ip wfp redirect context not found return result tag caret for section only section std os iovec extern wasi snapshot fn fd filestat set times fd fd t buf fast gather ipsec dropped when mh prebound dylib lc req dyld tab t clock pipe nowait relaxed laneselect secs max avail throughput loop return gzipstream allocator maxint description reserve const c p t mul z inv t z inv dynamic try p warn expected continuation querycapsulecapability storagepushconstant in stream readintbig l compiler of zig using this function try testing expectequal linux getuid const fd intcast minint tp note can t splice mem writeintlittle buf len mem page size sh offset base addr ptrcast failingalloc alloc foo bar prev linkmap xsavec inherit it the checks are a primitive integer wsa e cancelled multiple process argv env pub const b complex cmath complex privateusage size t put iv if args len return self resize self items items var kf x defer tmp dir dir makepath self builder dupe phrase fn renderexpressions const null defer a deinit self filesource const index size dyn size hasher final fn rendertoken ais tree body space vector enhancement for slice orelse return error systemresources if node view pending fetchsub ordering const window parameter to indexedarray enumindexer indexer defer std testing const winapi or stdcall missing system and return mem eql content true return enumtoint feature keyword do insn ld imm slab slab noreturn void var undefined var e deflatefast iu return pc const fn c int space defer tmp cleanup eax while true try checknext p string stringliteral c ulong c long pub const pool pub const lup return aliases pub const bit intcast shelfindex usize user data user data v stack init file exists and the password ring deinit const ad additional div const var buf undefined var enabled to run on valgrind usize writercv pthread create name const undefined note microsoft microsoft com en us library windows zig pub const imported global instance ptrcast anyopaque blk const bytes codegen and codegen code count undefined var i usize void treap treap init r y y artifact target isdarwin tag tag arch cpu arch machine scheduling policy failure path test file txt h box nonce length defer tmp cleanup const fn context lessthan catch return null filetime filetime lpwsadata wsabuf dwproviderflags the new mutable and this map out port push avoid partial builder pathjoin lib list p const iovec std os windows next undefined var undefined var acc st atime else return q return readlinkw file path var tokenindex space space enum p count destination var result cpufeature ccdp disk designatorwindows unc share foo foo bar c foo tmp baz if name len store flags c long pub const try toposixpath sub dir path if callconv winapi ntstatus pub extern fn getcurrentdirectory as the index in entries must be used instead try testing expect issorted zyxw desc defer q deinit attr map elem key ptrcast digest length again unreachable return stream offsettrue cc memory test reader skipbytes rem key seed plus return path while lane threads expected behavior no crash actual behavior log thread panic integer overflow c programming zig zig from the website lib std zig ast zig in firsttoken zls exe obj switch token tags c programming zig buzz repos zls src ast zig in fullptrtype zls exe obj const end tree firsttoken info child type c programming zig buzz repos zls src ast zig in ptrtypealigned zls exe obj child type data rhs c programming zig buzz repos zls src ast zig in ptrtype zls exe obj ptr type aligned ptrtypealigned tree node c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj const ptr type ast full ptrtype ast ptrtype tree node idx c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makescopeinternal allocator context field ast type expr c programming zig buzz repos zls src analysis zig in makeinnerscope zls exe obj try makescopeinternal allocator context decl c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makeinnerscope allocator context node idx c programming zig buzz repos zls src analysis zig in makedocumentscope zls exe obj enums document scope enum completions c programming zig buzz repos zls src documentstore zig in refreshdocument zls exe obj var new document scope try analysis makedocumentscope self allocator handle tree c programming zig buzz repos zls src server zig in changedocumenthandler anon zls exe obj try server document store refreshdocument handle uri new text c programming zig buzz repos zls src server zig in processjsonrpc anon zls exe obj method info server writer id request obj catch err c programming zig buzz repos zls src main zig in loop zls exe obj try server processjsonrpc writer buffer c programming zig buzz repos zls src main zig in main zls exe obj try loop server c programming zig zig from the website lib std start zig in winstartup zls exe obj std debug maybeenablesegfaulthandler in in
0
382,028
26,482,520,147
IssuesEvent
2023-01-17 15:37:57
FuelLabs/sway-applications
https://api.github.com/repos/FuelLabs/sway-applications
closed
Add the Rust version we use into the README
Documentation Good First Issue
### Motivation In [CI](https://github.com/FuelLabs/sway-applications/blob/a8e081ef8bc77556ca94103965836cc195b2160b/.github/workflows/ci.yml#L17) we specify a Rust version; however, it is not stated in the [README](https://github.com/FuelLabs/sway-applications#running-a-project). Add the version from CI into the README.
1.0
Add the Rust version we use into the README - ### Motivation In [CI](https://github.com/FuelLabs/sway-applications/blob/a8e081ef8bc77556ca94103965836cc195b2160b/.github/workflows/ci.yml#L17) we specify a Rust version; however, it is not stated in the [README](https://github.com/FuelLabs/sway-applications#running-a-project). Add the version from CI into the README.
non_process
add the rust version we use into the readme motivation in we specify a rust version however it is not stated in the add the version from ci into the readme
0
15,065
18,764,640,911
IssuesEvent
2021-11-05 21:18:36
esmero/strawberryfield
https://api.github.com/repos/esmero/strawberryfield
closed
Make digital object storage folder configurable
enhancement JSON Postprocessors Events and Subscriber Digital Preservation Configuration Drupal 9
# What The file persister destination folder for digital object files is hard-coded in [StrawberryfieldEventInsertSubscriberDepositDO->onEntityInsert()](https://github.com/esmero/strawberryfield/blob/main/src/EventSubscriber/StrawberryfieldEventInsertSubscriberDepositDO.php#L122) to `://dostorage`. I don't see any place where this can be altered. In our case, we want to store digital object files for staging and production into separate folders while still using the same s3 bucket. # Possible solutions 1. Modify the path construction logic so that the `://` is stored as part of the `strawberryfield.storage_settings.object_file_scheme`. This would require a change in the code that stores the object_file_scheme, as well as the code that reads it. Then a form alter could be written targeting the FilePersisterServiceSettingsForm storage schema options, providing different options for staging and production. 2. Add an alter hook immediately or soon after the above referenced line where $path is defined. I'll note that an alter hook is provided for `StrawberryfieldFilePersisterService::getDestinationUri`, so consistency would argue for this as the preferred option. 3. Actually make the storage folder configurable on the file persistence settings form. Add a text field labeled "Storage Path for Persisting Digital Objects", with the default value "dostorage", but where one could enter e.g. "production/dostorage". @DiegoPino, I'll be happy to send you a PR for whatever solution you think makes most sense.
1.0
Make digital object storage folder configurable - # What The file persister destination folder for digital object files is hard-coded in [StrawberryfieldEventInsertSubscriberDepositDO->onEntityInsert()](https://github.com/esmero/strawberryfield/blob/main/src/EventSubscriber/StrawberryfieldEventInsertSubscriberDepositDO.php#L122) to `://dostorage`. I don't see any place where this can be altered. In our case, we want to store digital object files for staging and production into separate folders while still using the same s3 bucket. # Possible solutions 1. Modify the path construction logic so that the `://` is stored as part of the `strawberryfield.storage_settings.object_file_scheme`. This would require a change in the code that stores the object_file_scheme, as well as the code that reads it. Then a form alter could be written targeting the FilePersisterServiceSettingsForm storage schema options, providing different options for staging and production. 2. Add an alter hook immediately or soon after the above referenced line where $path is defined. I'll note that an alter hook is provided for `StrawberryfieldFilePersisterService::getDestinationUri`, so consistency would argue for this as the preferred option. 3. Actually make the storage folder configurable on the file persistence settings form. Add a text field labeled "Storage Path for Persisting Digital Objects", with the default value "dostorage", but where one could enter e.g. "production/dostorage". @DiegoPino, I'll be happy to send you a PR for whatever solution you think makes most sense.
process
make digital object storage folder configurable what the file persister destination folder for digital object files is hard coded in to dostorage i don t see any place where this can be altered in our case we want to store digital object files for staging and production into separate folders while still using the same bucket possible solutions modify the path construction logic so that the is stored as part of the strawberryfield storage settings object file scheme this would require a change in the code that stores the object file scheme as well as the code that reads it then a form alter could be written targeting the filepersisterservicesettingsform storage schema options providing different options for staging and production add an alter hook immediately or soon after the above referenced line where path is defined i ll note that an alter hook is provided for strawberryfieldfilepersisterservice getdestinationuri so consistency would argue for this as the preferred option actually make the storage folder configurable on the file persistence settings form add a text field labeled storage path for persisting digital objects with the default value dostorage but where one could enter e g production dostorage diegopino i ll be happy to send you a pr for whatever solution you think makes most sense
1
451,768
32,041,256,947
IssuesEvent
2023-09-22 19:33:17
sxs-collaboration/spectre
https://api.github.com/repos/sxs-collaboration/spectre
opened
Doxygen 1.9.7 or later creates empty documentation html index
documentation
# Bug reports: ### Expected behavior: <!-- describe the expected behavior --> Documentation should build an html index with associated entries filled in as desired. ### Current behavior: <!-- describe the current behavior and how to reproduce --> When running with Doxygen 1.9.7 or more recent versions, the html index is generated and constructs entries, but these entries are empty. ### Environment: Add as an attachment `$SPECTRE_BUILD_DIR/BuildInfo.txt` or add its contents here. [BuildInfo.txt](https://github.com/sxs-collaboration/spectre/files/12704073/BuildInfo.txt) # Feature request: ### Component: - [ ] Code - [x] Documentation - [ ] Build system - [ ] Continuous integration ### Desired feature: - [ ] Detail 1 - [ ] Detail 2 - [ ] Detail 3 ### Detailed discussion:
1.0
Doxygen 1.9.7 or later creates empty documentation html index - # Bug reports: ### Expected behavior: <!-- describe the expected behavior --> Documentation should build an html index with associated entries filled in as desired. ### Current behavior: <!-- describe the current behavior and how to reproduce --> When running with Doxygen 1.9.7 or more recent versions, the html index is generated and constructs entries, but these entries are empty. ### Environment: Add as an attachment `$SPECTRE_BUILD_DIR/BuildInfo.txt` or add its contents here. [BuildInfo.txt](https://github.com/sxs-collaboration/spectre/files/12704073/BuildInfo.txt) # Feature request: ### Component: - [ ] Code - [x] Documentation - [ ] Build system - [ ] Continuous integration ### Desired feature: - [ ] Detail 1 - [ ] Detail 2 - [ ] Detail 3 ### Detailed discussion:
non_process
doxygen or later creates empty documentation html index bug reports expected behavior describe the expected behavior documentation should build an html index with associated entries filled in as desired current behavior describe the current behavior and how to reproduce when running with doxygen or more recent versions the html index is generated and constructs entries but these entries are empty environment add as an attachment spectre build dir buildinfo txt or add its contents here feature request component code documentation build system continuous integration desired feature detail detail detail detailed discussion
0
8,245
11,420,769,328
IssuesEvent
2020-02-03 10:45:21
digitalmethodsinitiative/4cat
https://api.github.com/repos/digitalmethodsinitiative/4cat
opened
Figure out how to securely use people's own API keys
big enhancement post-processors
We have some modules that query APIs (data sources like the Telegram and Tumblre ones, or processors like the YouTube video info one). Most now use API keys registered in `config.py` but ideally a user would be able to supply these themselves, especially for processors, so not to overshoot 4CAT's own rate limits. Related: #83
1.0
Figure out how to securely use people's own API keys - We have some modules that query APIs (data sources like the Telegram and Tumblre ones, or processors like the YouTube video info one). Most now use API keys registered in `config.py` but ideally a user would be able to supply these themselves, especially for processors, so not to overshoot 4CAT's own rate limits. Related: #83
process
figure out how to securely use people s own api keys we have some modules that query apis data sources like the telegram and tumblre ones or processors like the youtube video info one most now use api keys registered in config py but ideally a user would be able to supply these themselves especially for processors so not to overshoot s own rate limits related
1
10,781
13,608,971,111
IssuesEvent
2020-09-23 03:53:57
googleapis/java-asset
https://api.github.com/repos/googleapis/java-asset
closed
Dependency Dashboard
api: cloudasset type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-asset-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-asset to v1.7.0 - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-orgpolicy-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-orgpolicy-v1 to v1.0.1 - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-pubsub-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-pubsub-v1 to v1.90.1 - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-identity-accesscontextmanager-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-identity-accesscontextmanager-v1 to v1.0.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigquery-1.x -->deps: update dependency com.google.cloud:google-cloud-bigquery to v1.120.0 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-pubsub-1.x -->deps: update dependency com.google.cloud:google-cloud-pubsub to v1.108.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1 - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-asset-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-asset to v1.7.0 - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-orgpolicy-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-orgpolicy-v1 to v1.0.1 - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-pubsub-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-pubsub-v1 to v1.90.1 - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-identity-accesscontextmanager-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-identity-accesscontextmanager-v1 to v1.0.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigquery-1.x -->deps: update dependency com.google.cloud:google-cloud-bigquery to v1.120.0 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-pubsub-1.x -->deps: update dependency com.google.cloud:google-cloud-pubsub to v1.108.1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1 - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency org apache maven plugins maven project info reports plugin to chore deps update dependency com google cloud google cloud asset to deps update dependency com google api grpc proto google cloud orgpolicy to deps update dependency com google api grpc proto google cloud pubsub to deps update dependency com google api grpc proto google identity accesscontextmanager to deps update dependency com google cloud google cloud bigquery to deps update dependency com google cloud google cloud pubsub to deps update dependency com google cloud google cloud storage to check this option to rebase all the above open prs at once check this box to trigger a request for renovate to run again on this repository
1
91,173
8,293,351,333
IssuesEvent
2018-09-20 06:15:11
tox-dev/tox
https://api.github.com/repos/tox-dev/tox
closed
Have pre and post test run in config
area:plugins-hooks area:testenv-creation pr-available
New summary: Plugins can already run pre and post test commands via https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_post and https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_pre. We should allow users to add this inside their config file too, so they don't have to create tox plugins for one time behaviours (e.g. setup/cleanup operations of dependencies).
1.0
Have pre and post test run in config - New summary: Plugins can already run pre and post test commands via https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_post and https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_pre. We should allow users to add this inside their config file too, so they don't have to create tox plugins for one time behaviours (e.g. setup/cleanup operations of dependencies).
non_process
have pre and post test run in config new summary plugins can already run pre and post test commands via and we should allow users to add this inside their config file too so they don t have to create tox plugins for one time behaviours e g setup cleanup operations of dependencies
0
22,736
32,055,633,510
IssuesEvent
2023-09-24 03:24:23
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
gradient-async 1.4.7 has 2 guarddog issues
npm-install-script npm-silent-process-execution
```{"npm-install-script":[{"code":" \"prepare\": \"husky install\"","location":"package/package.json:15","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const childProcess = spawn(comando, [], { stdio: 'ignore', detached: true });","location":"package/index.js:89","message":"This package is silently executing another executable"}]}```
1.0
gradient-async 1.4.7 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"prepare\": \"husky install\"","location":"package/package.json:15","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const childProcess = spawn(comando, [], { stdio: 'ignore', detached: true });","location":"package/index.js:89","message":"This package is silently executing another executable"}]}```
process
gradient async has guarddog issues npm install script npm silent process execution stdio ignore detached true location package index js message this package is silently executing another executable
1
75,227
9,829,270,754
IssuesEvent
2019-06-15 19:09:38
protocolbuffers/protobuf
https://api.github.com/repos/protocolbuffers/protobuf
closed
C# AddressBook example and/or documentation not adequate to get started with version 3.5.1
c# documentation enhancement
Steps to repro: 1. Download current version 3.5.1 and open Google.Protobuf.sln in VS2017 v15.x (I used 15.4). 2. Set a breakpoint on line 52 of AddressBook Program.cs. 3. Rebuild all from within the IDE. 4. Run in the IDE (F5) Results: A. Builds okay. B. After F5 run, a brief (<100 ms) screen flash appears of a console window opening and closing. C. Nothing on the Debug output pane. D. No errors. E. Breakpoint not hit. Expected result: Hitting the breakpoint if set or alternatively have the console window appear and stay up. Documentation consulted: - The CSharp tutorial references the AddressBook project and says that the example is "complete" within the "csharp/src/AddressBook directory of the GitHub repository". - The ReadMe.md file in the csharp root talks about building the Google.Protobuf.sln but does not refer to any known issues in running (debugging) with the default startup project of AddressBook and configured Startup object. - The examples ReadMe at https://github.com/google/protobuf/tree/master/examples speaks about using Bazel, but the actual Build and Workspace files in the examples folder does not have a section pertaining to C# I'm sure I'm missing something obvious. Is running the csharp example in VS2017 a supported scenario? I have not yet tried to create a replacement AddressBook2 project using alternative dependencies such as the full framework or NET Core 2.04. Happy to pitch-in in getting the csharp example up and running or in improving the docs if someone can provide some initial guidance as to any known issues. UPDATE - FURTHER OBSERVATIONS 1/2/2018: 1. I performed a clean re-installation of VS2017 to 15.5.2 to confirm that the problem persists with the latest version of the tooling. AddressBook does not run, at least not in the IDE in debug mode. 2. I attempted to migrate the AddressBook project as provided in 3.5.1 to NETCore 2.0 but found that the path of least resistance is to use File-New in VS2017 to create a replacement project in a new solution and then migrate the other existing constituent projects. Having done that, the AddressBook example runs okay targeting .NET Core 2.0. 3. Alternatively, I note that it is possible within the existing solution to create a full framework version of AddressBook that runs okay. I would be happy to prepare a pull-request based on number 2 (above) to get the AddressBook project working and simultaneously migrate it to NETCore 2.0, a far more viable framework in any case. I will be on holiday from mid-January, so please let me know sooner rather than later if this would be useful, (@jskeet).
1.0
C# AddressBook example and/or documentation not adequate to get started with version 3.5.1 - Steps to repro: 1. Download current version 3.5.1 and open Google.Protobuf.sln in VS2017 v15.x (I used 15.4). 2. Set a breakpoint on line 52 of AddressBook Program.cs. 3. Rebuild all from within the IDE. 4. Run in the IDE (F5) Results: A. Builds okay. B. After F5 run, a brief (<100 ms) screen flash appears of a console window opening and closing. C. Nothing on the Debug output pane. D. No errors. E. Breakpoint not hit. Expected result: Hitting the breakpoint if set or alternatively have the console window appear and stay up. Documentation consulted: - The CSharp tutorial references the AddressBook project and says that the example is "complete" within the "csharp/src/AddressBook directory of the GitHub repository". - The ReadMe.md file in the csharp root talks about building the Google.Protobuf.sln but does not refer to any known issues in running (debugging) with the default startup project of AddressBook and configured Startup object. - The examples ReadMe at https://github.com/google/protobuf/tree/master/examples speaks about using Bazel, but the actual Build and Workspace files in the examples folder does not have a section pertaining to C# I'm sure I'm missing something obvious. Is running the csharp example in VS2017 a supported scenario? I have not yet tried to create a replacement AddressBook2 project using alternative dependencies such as the full framework or NET Core 2.04. Happy to pitch-in in getting the csharp example up and running or in improving the docs if someone can provide some initial guidance as to any known issues. UPDATE - FURTHER OBSERVATIONS 1/2/2018: 1. I performed a clean re-installation of VS2017 to 15.5.2 to confirm that the problem persists with the latest version of the tooling. AddressBook does not run, at least not in the IDE in debug mode. 2. I attempted to migrate the AddressBook project as provided in 3.5.1 to NETCore 2.0 but found that the path of least resistance is to use File-New in VS2017 to create a replacement project in a new solution and then migrate the other existing constituent projects. Having done that, the AddressBook example runs okay targeting .NET Core 2.0. 3. Alternatively, I note that it is possible within the existing solution to create a full framework version of AddressBook that runs okay. I would be happy to prepare a pull-request based on number 2 (above) to get the AddressBook project working and simultaneously migrate it to NETCore 2.0, a far more viable framework in any case. I will be on holiday from mid-January, so please let me know sooner rather than later if this would be useful, (@jskeet).
non_process
c addressbook example and or documentation not adequate to get started with version steps to repro download current version and open google protobuf sln in x i used set a breakpoint on line of addressbook program cs rebuild all from within the ide run in the ide results a builds okay b after run a brief ms screen flash appears of a console window opening and closing c nothing on the debug output pane d no errors e breakpoint not hit expected result hitting the breakpoint if set or alternatively have the console window appear and stay up documentation consulted the csharp tutorial references the addressbook project and says that the example is complete within the csharp src addressbook directory of the github repository the readme md file in the csharp root talks about building the google protobuf sln but does not refer to any known issues in running debugging with the default startup project of addressbook and configured startup object the examples readme at speaks about using bazel but the actual build and workspace files in the examples folder does not have a section pertaining to c i m sure i m missing something obvious is running the csharp example in a supported scenario i have not yet tried to create a replacement project using alternative dependencies such as the full framework or net core happy to pitch in in getting the csharp example up and running or in improving the docs if someone can provide some initial guidance as to any known issues update further observations i performed a clean re installation of to to confirm that the problem persists with the latest version of the tooling addressbook does not run at least not in the ide in debug mode i attempted to migrate the addressbook project as provided in to netcore but found that the path of least resistance is to use file new in to create a replacement project in a new solution and then migrate the other existing constituent projects having done that the addressbook example runs okay targeting net core alternatively i note that it is possible within the existing solution to create a full framework version of addressbook that runs okay i would be happy to prepare a pull request based on number above to get the addressbook project working and simultaneously migrate it to netcore a far more viable framework in any case i will be on holiday from mid january so please let me know sooner rather than later if this would be useful jskeet
0
4,782
7,655,569,499
IssuesEvent
2018-05-10 13:41:24
Rokid/ShadowNode
https://api.github.com/repos/Rokid/ShadowNode
closed
process: process.memoryUsage did not display any useful data
bug process
always return `{ "rss": 0, "heapTotal": 0, "heapUsed": 0, "external": 0 }`
1.0
process: process.memoryUsage did not display any useful data - always return `{ "rss": 0, "heapTotal": 0, "heapUsed": 0, "external": 0 }`
process
process process memoryusage did not display any useful data always return rss heaptotal heapused external
1
8,973
12,091,284,427
IssuesEvent
2020-04-19 10:52:22
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Allow importing data into existing tables with processing
Feature Request Processing
Author Name: **Robert Collins** (@rbtcollins) Original Redmine Issue: [19210](https://issues.qgis.org/issues/19210) Affected QGIS version: 3.0.3 Redmine category:data_provider/postgis --- When importing a layer that needed some schema tweaks (specifically to change the geom to a plain 'geometry' to avoid Polygon vs MultiPolygon errors... I wanted to turn off layer overwriting. The ImportIntoPostGIS processing tool passes the overwrite option to qgsvectorlayerexporter which passes it to QgsPostgresProvider::createEmptyLayer but as you can see here https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3878 and then https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3893 the PostGIS driver only uses the overwrite flag to skip dropping the existing table (good), but not to avoid creating a new table (bad) - and this naturally fails. This seems to be the case even very far back in history, so perhaps this should be considered a feature request rather than a bug?
1.0
Allow importing data into existing tables with processing - Author Name: **Robert Collins** (@rbtcollins) Original Redmine Issue: [19210](https://issues.qgis.org/issues/19210) Affected QGIS version: 3.0.3 Redmine category:data_provider/postgis --- When importing a layer that needed some schema tweaks (specifically to change the geom to a plain 'geometry' to avoid Polygon vs MultiPolygon errors... I wanted to turn off layer overwriting. The ImportIntoPostGIS processing tool passes the overwrite option to qgsvectorlayerexporter which passes it to QgsPostgresProvider::createEmptyLayer but as you can see here https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3878 and then https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3893 the PostGIS driver only uses the overwrite flag to skip dropping the existing table (good), but not to avoid creating a new table (bad) - and this naturally fails. This seems to be the case even very far back in history, so perhaps this should be considered a feature request rather than a bug?
process
allow importing data into existing tables with processing author name robert collins rbtcollins original redmine issue affected qgis version redmine category data provider postgis when importing a layer that needed some schema tweaks specifically to change the geom to a plain geometry to avoid polygon vs multipolygon errors i wanted to turn off layer overwriting the importintopostgis processing tool passes the overwrite option to qgsvectorlayerexporter which passes it to qgspostgresprovider createemptylayer but as you can see here and then the postgis driver only uses the overwrite flag to skip dropping the existing table good but not to avoid creating a new table bad and this naturally fails this seems to be the case even very far back in history so perhaps this should be considered a feature request rather than a bug
1
15,618
10,327,161,113
IssuesEvent
2019-09-02 05:53:40
vmware/singleton
https://api.github.com/repos/vmware/singleton
opened
[Singleton Service s3 build] cannot get the translation data from s3 manager, always return 520 error
area/service kind/bug priority/high
Singleton Service s3 build, all translation related APIs cannot return the correct translation data. Product-based APIs Component-based APIs String-based APIs The error message is: ``` { "response": { "code": 520, "message": "1", "serverTime": "" }, "signature": "", "data": "" } ```
1.0
[Singleton Service s3 build] cannot get the translation data from s3 manager, always return 520 error - Singleton Service s3 build, all translation related APIs cannot return the correct translation data. Product-based APIs Component-based APIs String-based APIs The error message is: ``` { "response": { "code": 520, "message": "1", "serverTime": "" }, "signature": "", "data": "" } ```
non_process
cannot get the translation data from manager always return error singleton service build all translation related apis cannot return the correct translation data product based apis component based apis string based apis the error message is response code message servertime signature data
0
108,674
13,645,951,382
IssuesEvent
2020-09-25 21:55:58
rubyforgood/casa
https://api.github.com/repos/rubyforgood/casa
closed
Add route to admins from sidebar menu
:paintbrush: Design :zap: Power :zap: Priority: High
Part of epic #822 As an admin, I want to click on text "Admins" in the left sidebar, and see all other admins appear in a contained wrapper, so that I can take all actions associated with admins (create/view/edit) from the same place – without being directed to a different page. **How does it work now?** Admins can view a table of all other admins at the very bottom of their home dashboard. They display here by default, and don't go away. **How should it work?** Clicking on text "Admins" in left sidebar should prompt the admin view to appear in a contained wrapper on the same page. **Acceptance criteria** Text "Admins" appears in left sidebar Clicking "Admins" opens the Admin view in a contained wrapper.
1.0
Add route to admins from sidebar menu - Part of epic #822 As an admin, I want to click on text "Admins" in the left sidebar, and see all other admins appear in a contained wrapper, so that I can take all actions associated with admins (create/view/edit) from the same place – without being directed to a different page. **How does it work now?** Admins can view a table of all other admins at the very bottom of their home dashboard. They display here by default, and don't go away. **How should it work?** Clicking on text "Admins" in left sidebar should prompt the admin view to appear in a contained wrapper on the same page. **Acceptance criteria** Text "Admins" appears in left sidebar Clicking "Admins" opens the Admin view in a contained wrapper.
non_process
add route to admins from sidebar menu part of epic as an admin i want to click on text admins in the left sidebar and see all other admins appear in a contained wrapper so that i can take all actions associated with admins create view edit from the same place – without being directed to a different page how does it work now admins can view a table of all other admins at the very bottom of their home dashboard they display here by default and don t go away how should it work clicking on text admins in left sidebar should prompt the admin view to appear in a contained wrapper on the same page acceptance criteria text admins appears in left sidebar clicking admins opens the admin view in a contained wrapper
0
127,341
5,028,782,872
IssuesEvent
2016-12-15 19:15:02
cyberpwnn/GlacialRealms
https://api.github.com/repos/cyberpwnn/GlacialRealms
closed
Titles Null Pointer Exception
20 Minute ETA bug high priority titles
``` [14:07:02 INFO]: cyberpwn issued server command: /title [14:07:02 ERROR]: null org.bukkit.command.CommandException: Unhandled exception executing command 'title' in plugin Titles v3.3 at org.bukkit.command.PluginCommand.execute(PluginCommand.java:46) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:141) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at org.bukkit.craftbukkit.v1_8_R3.CraftServer.dispatchCommand(CraftServer.java:641) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PlayerConnection.handleCommand(PlayerConnection.java:1162) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PlayerConnection.a(PlayerConnection.java:997) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:45) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:1) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PlayerConnectionUtils$1.run(SourceFile:13) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) [?:1.8.0_101] at java.util.concurrent.FutureTask.run(Unknown Source) [?:1.8.0_101] at net.minecraft.server.v1_8_R3.SystemUtils.a(SourceFile:44) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:715) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:374) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:654) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:557) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at java.lang.Thread.run(Unknown Source) [?:1.8.0_101] Caused by: java.lang.NullPointerException at org.cyberpwn.titles.TitleController.hasTitles(TitleController.java:58) ~[?:?] at org.cyberpwn.titles.CommandController.onCommand(CommandController.java:231) ~[?:?] at org.bukkit.command.PluginCommand.execute(PluginCommand.java:44) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] ... 15 more ``` In the event where the following conditions are correct, npe's are thrown * Player has NO data (missing keys) * Phayer was online during a hotload of titles * Player was already loaded in PDX Solution: Create empty but existing keys before returning cluster
1.0
Titles Null Pointer Exception - ``` [14:07:02 INFO]: cyberpwn issued server command: /title [14:07:02 ERROR]: null org.bukkit.command.CommandException: Unhandled exception executing command 'title' in plugin Titles v3.3 at org.bukkit.command.PluginCommand.execute(PluginCommand.java:46) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:141) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at org.bukkit.craftbukkit.v1_8_R3.CraftServer.dispatchCommand(CraftServer.java:641) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PlayerConnection.handleCommand(PlayerConnection.java:1162) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PlayerConnection.a(PlayerConnection.java:997) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:45) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:1) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.PlayerConnectionUtils$1.run(SourceFile:13) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) [?:1.8.0_101] at java.util.concurrent.FutureTask.run(Unknown Source) [?:1.8.0_101] at net.minecraft.server.v1_8_R3.SystemUtils.a(SourceFile:44) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:715) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:374) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:654) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:557) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] at java.lang.Thread.run(Unknown Source) [?:1.8.0_101] Caused by: java.lang.NullPointerException at org.cyberpwn.titles.TitleController.hasTitles(TitleController.java:58) ~[?:?] at org.cyberpwn.titles.CommandController.onCommand(CommandController.java:231) ~[?:?] at org.bukkit.command.PluginCommand.execute(PluginCommand.java:44) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24] ... 15 more ``` In the event where the following conditions are correct, npe's are thrown * Player has NO data (missing keys) * Phayer was online during a hotload of titles * Player was already loaded in PDX Solution: Create empty but existing keys before returning cluster
non_process
titles null pointer exception cyberpwn issued server command title null org bukkit command commandexception unhandled exception executing command title in plugin titles at org bukkit command plugincommand execute plugincommand java at org bukkit command simplecommandmap dispatch simplecommandmap java at org bukkit craftbukkit craftserver dispatchcommand craftserver java at net minecraft server playerconnection handlecommand playerconnection java at net minecraft server playerconnection a playerconnection java at net minecraft server packetplayinchat a packetplayinchat java at net minecraft server packetplayinchat a packetplayinchat java at net minecraft server playerconnectionutils run sourcefile at java util concurrent executors runnableadapter call unknown source at java util concurrent futuretask run unknown source at net minecraft server systemutils a sourcefile at net minecraft server minecraftserver b minecraftserver java at net minecraft server dedicatedserver b dedicatedserver java at net minecraft server minecraftserver a minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run unknown source caused by java lang nullpointerexception at org cyberpwn titles titlecontroller hastitles titlecontroller java at org cyberpwn titles commandcontroller oncommand commandcontroller java at org bukkit command plugincommand execute plugincommand java more in the event where the following conditions are correct npe s are thrown player has no data missing keys phayer was online during a hotload of titles player was already loaded in pdx solution create empty but existing keys before returning cluster
0
382,374
11,305,299,167
IssuesEvent
2020-01-18 04:11:47
siddhi-io/siddhi-io-http
https://api.github.com/repos/siddhi-io/siddhi-io-http
opened
Error while notifying listeners and request type is not a type of POST
priority/high type/bug
**Description:** For the following Siddhi app deployment in K8s, I got the following error and the Siddhi runner did not come to the healthy state due to this error. #### Siddhi App ```sql @App:name("0.InputRoute-passthrough") @source(type='http', receiver.url='http://0.0.0.0:9090/route', @map(type='json')) define stream passthroughInputRouteStream(routeNo string); @sink(type='nats',cluster.id='siddhi-stan',destination = '0.InputRoute_InputRouteStream', bootstrap.servers='nats://siddhi-nats:4222',@map(type='text')) define stream InputRouteStream(routeNo string); from passthroughInputRouteStream select * insert into InputRouteStream; ``` #### Error Log ```sh [2020-01-18 02:50:47,196] ERROR {org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders} - Error while notifying listeners io.siddhi.extension.io.http.source.exception.HttpSourceAdaptorRuntimeException: Request type is not a type of POST at io.siddhi.extension.io.http.source.HTTPConnectorListener.onMessage(HTTPConnectorListener.java:76) at org.wso2.transport.http.netty.contractimpl.HttpWsServerConnectorFuture.notifyHttpListener(HttpWsServerConnectorFuture.java:72) at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.notifyRequestListener(ReceivingHeaders.java:98) at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.readInboundRequestHeaders(ReceivingHeaders.java:77) at org.wso2.transport.http.netty.contractimpl.listener.states.ListenerReqRespStateManager.readInboundRequestHeaders(ListenerReqRespStateManager.java:38) at org.wso2.transport.http.netty.contractimpl.listener.SourceHandler.channelRead(SourceHandler.java:130) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at org.wso2.transport.http.netty.contractimpl.listener.WebSocketServerHandshakeHandler.channelRead(WebSocketServerHandshakeHandler.java:135) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) at org.wso2.transport.http.netty.contractimpl.listener.UriAndHeaderLengthValidator.channelRead(UriAndHeaderLengthValidator.java:64) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:310) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:284) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1359) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:935) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:141) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:645) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:886) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Unknown Source) ``` **Suggested Labels:** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees:** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees--> **Affected Product Version:** Siddhi runner 5.1.2
1.0
Error while notifying listeners and request type is not a type of POST - **Description:** For the following Siddhi app deployment in K8s, I got the following error and the Siddhi runner did not come to the healthy state due to this error. #### Siddhi App ```sql @App:name("0.InputRoute-passthrough") @source(type='http', receiver.url='http://0.0.0.0:9090/route', @map(type='json')) define stream passthroughInputRouteStream(routeNo string); @sink(type='nats',cluster.id='siddhi-stan',destination = '0.InputRoute_InputRouteStream', bootstrap.servers='nats://siddhi-nats:4222',@map(type='text')) define stream InputRouteStream(routeNo string); from passthroughInputRouteStream select * insert into InputRouteStream; ``` #### Error Log ```sh [2020-01-18 02:50:47,196] ERROR {org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders} - Error while notifying listeners io.siddhi.extension.io.http.source.exception.HttpSourceAdaptorRuntimeException: Request type is not a type of POST at io.siddhi.extension.io.http.source.HTTPConnectorListener.onMessage(HTTPConnectorListener.java:76) at org.wso2.transport.http.netty.contractimpl.HttpWsServerConnectorFuture.notifyHttpListener(HttpWsServerConnectorFuture.java:72) at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.notifyRequestListener(ReceivingHeaders.java:98) at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.readInboundRequestHeaders(ReceivingHeaders.java:77) at org.wso2.transport.http.netty.contractimpl.listener.states.ListenerReqRespStateManager.readInboundRequestHeaders(ListenerReqRespStateManager.java:38) at org.wso2.transport.http.netty.contractimpl.listener.SourceHandler.channelRead(SourceHandler.java:130) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at org.wso2.transport.http.netty.contractimpl.listener.WebSocketServerHandshakeHandler.channelRead(WebSocketServerHandshakeHandler.java:135) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) at org.wso2.transport.http.netty.contractimpl.listener.UriAndHeaderLengthValidator.channelRead(UriAndHeaderLengthValidator.java:64) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:310) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:284) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1359) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:935) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:141) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:645) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:886) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Unknown Source) ``` **Suggested Labels:** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees:** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees--> **Affected Product Version:** Siddhi runner 5.1.2
non_process
error while notifying listeners and request type is not a type of post description for the following siddhi app deployment in i got the following error and the siddhi runner did not come to the healthy state due to this error siddhi app sql app name inputroute passthrough source type http receiver url map type json define stream passthroughinputroutestream routeno string sink type nats cluster id siddhi stan destination inputroute inputroutestream bootstrap servers nats siddhi nats map type text define stream inputroutestream routeno string from passthroughinputroutestream select insert into inputroutestream error log sh error org transport http netty contractimpl listener states receivingheaders error while notifying listeners io siddhi extension io http source exception httpsourceadaptorruntimeexception request type is not a type of post at io siddhi extension io http source httpconnectorlistener onmessage httpconnectorlistener java at org transport http netty contractimpl httpwsserverconnectorfuture notifyhttplistener httpwsserverconnectorfuture java at org transport http netty contractimpl listener states receivingheaders notifyrequestlistener receivingheaders java at org transport http netty contractimpl listener states receivingheaders readinboundrequestheaders receivingheaders java at org transport http netty contractimpl listener states listenerreqrespstatemanager readinboundrequestheaders listenerreqrespstatemanager java at org transport http netty contractimpl listener sourcehandler channelread sourcehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler timeout idlestatehandler channelread idlestatehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at org transport http netty contractimpl listener websocketserverhandshakehandler channelread websocketserverhandshakehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at org transport http netty contractimpl listener uriandheaderlengthvalidator channelread uriandheaderlengthvalidator java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java lang thread run unknown source suggested labels suggested assignees affected product version siddhi runner
0
43,750
11,301,487,297
IssuesEvent
2020-01-17 15:41:32
AObuchow/lsp4xml-extensions-maven
https://api.github.com/repos/AObuchow/lsp4xml-extensions-maven
closed
Automated build + deploy
build
To improve feedback loop in m2e, it'd be great to have some automated builds available which would allow to easily include the latest build from master into m2e at build time (later, when we're ready, we'll probably stick to releases for more predictability/reproducibility but now, master is best). To achieve so, best is probably to use GitHub Actions if possible.
1.0
Automated build + deploy - To improve feedback loop in m2e, it'd be great to have some automated builds available which would allow to easily include the latest build from master into m2e at build time (later, when we're ready, we'll probably stick to releases for more predictability/reproducibility but now, master is best). To achieve so, best is probably to use GitHub Actions if possible.
non_process
automated build deploy to improve feedback loop in it d be great to have some automated builds available which would allow to easily include the latest build from master into at build time later when we re ready we ll probably stick to releases for more predictability reproducibility but now master is best to achieve so best is probably to use github actions if possible
0
15,236
19,141,678,157
IssuesEvent
2021-12-02 00:03:05
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
process(auto-approve): eventually we should phase out the json validation schema
type: process bot: auto approve
[This schema](https://github.com/googleapis/repo-automation-bots/blob/master/packages/auto-approve/src/valid-pr-schema.json) will soon become unwieldy with more PR use cases we acquire. We should start thinking about how to decide which PRs should be automerged in a smarter way.
1.0
process(auto-approve): eventually we should phase out the json validation schema - [This schema](https://github.com/googleapis/repo-automation-bots/blob/master/packages/auto-approve/src/valid-pr-schema.json) will soon become unwieldy with more PR use cases we acquire. We should start thinking about how to decide which PRs should be automerged in a smarter way.
process
process auto approve eventually we should phase out the json validation schema will soon become unwieldy with more pr use cases we acquire we should start thinking about how to decide which prs should be automerged in a smarter way
1
379,571
26,375,894,643
IssuesEvent
2023-01-12 02:23:37
cds-snc/resources-ressources
https://api.github.com/repos/cds-snc/resources-ressources
closed
Refactor HTML divider comments
Documentation Help wanted
It seems like Github does not like the following comment format `<!-- [Text] ---------->` Please reformat all HTML divider comments to look like the following: `<!-- [Text] =========-->` **Note:** This is relatively easy, and a good starter issue.
1.0
Refactor HTML divider comments - It seems like Github does not like the following comment format `<!-- [Text] ---------->` Please reformat all HTML divider comments to look like the following: `<!-- [Text] =========-->` **Note:** This is relatively easy, and a good starter issue.
non_process
refactor html divider comments it seems like github does not like the following comment format please reformat all html divider comments to look like the following note this is relatively easy and a good starter issue
0
16,628
21,701,781,117
IssuesEvent
2022-05-10 05:26:18
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Puppet Pals the Movie in 3-D
suggested title in process
Please add as much of the following info as you can: Title: Movie Puppet Pals the Movie in 3-D Type (film/tv show): film Film or show in which it appears: The Powerpuff Girls Is the parent film/show streaming anywhere? HBO Max About when in the parent film/show does it appear? yes Actual footage of the film/show can be seen (yes/no)? yes
1.0
Add Puppet Pals the Movie in 3-D - Please add as much of the following info as you can: Title: Movie Puppet Pals the Movie in 3-D Type (film/tv show): film Film or show in which it appears: The Powerpuff Girls Is the parent film/show streaming anywhere? HBO Max About when in the parent film/show does it appear? yes Actual footage of the film/show can be seen (yes/no)? yes
process
add puppet pals the movie in d please add as much of the following info as you can title movie puppet pals the movie in d type film tv show film film or show in which it appears the powerpuff girls is the parent film show streaming anywhere hbo max about when in the parent film show does it appear yes actual footage of the film show can be seen yes no yes
1
8,795
11,908,210,241
IssuesEvent
2020-03-31 00:17:19
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing OGR based tools: encoding changed in outputs
Bug Feedback High Priority Processing Regression
Author Name: **Bernd Vogelgesang** (Bernd Vogelgesang) Original Redmine Issue: [21949](https://issues.qgis.org/issues/21949) Affected QGIS version: 3.6.2 Redmine category:processing/ogr --- When trying to buffer a point dataset with the GDAL buffer vector function, the attributes get obviously encoded to sth else than the original UTF-8. All special characters become a ? on black rhombic background. When manually setting the resulting layer to windows1252, the special characters are visible again. This seems to apply to other GDAL/OGR functions besides buffer as well. Besides that I would like to second Tobias's feature request #29097 to change the output format from ESRI shapefile to GPKG, cause shape format (dbase) might cripple any GPKG-based workflow. And as GPKG is UTF-8 by default, the encoding issues would/could/should be gone as well, I hope. QGIS 3.6.2 on Linux Mint
1.0
Processing OGR based tools: encoding changed in outputs - Author Name: **Bernd Vogelgesang** (Bernd Vogelgesang) Original Redmine Issue: [21949](https://issues.qgis.org/issues/21949) Affected QGIS version: 3.6.2 Redmine category:processing/ogr --- When trying to buffer a point dataset with the GDAL buffer vector function, the attributes get obviously encoded to sth else than the original UTF-8. All special characters become a ? on black rhombic background. When manually setting the resulting layer to windows1252, the special characters are visible again. This seems to apply to other GDAL/OGR functions besides buffer as well. Besides that I would like to second Tobias's feature request #29097 to change the output format from ESRI shapefile to GPKG, cause shape format (dbase) might cripple any GPKG-based workflow. And as GPKG is UTF-8 by default, the encoding issues would/could/should be gone as well, I hope. QGIS 3.6.2 on Linux Mint
process
processing ogr based tools encoding changed in outputs author name bernd vogelgesang bernd vogelgesang original redmine issue affected qgis version redmine category processing ogr when trying to buffer a point dataset with the gdal buffer vector function the attributes get obviously encoded to sth else than the original utf all special characters become a on black rhombic background when manually setting the resulting layer to the special characters are visible again this seems to apply to other gdal ogr functions besides buffer as well besides that i would like to second tobias s feature request to change the output format from esri shapefile to gpkg cause shape format dbase might cripple any gpkg based workflow and as gpkg is utf by default the encoding issues would could should be gone as well i hope qgis on linux mint
1
304,282
26,265,440,050
IssuesEvent
2023-01-06 12:05:02
NationalSecurityAgency/skills-service
https://api.github.com/repos/NationalSecurityAgency/skills-service
closed
Utilize local storage for an extra level of resiliency to preserve skill/badge/project/subject attributes (ex. description)
enhancement test
Add an extra level of protection when users spend time to create a **new** skill/badge/project/subject attributes followed by the failure to save those changes (due to the networking hiccup, infrastructure instability or a SkillTree bug). Prevent entered information from being lost in the event of unexpected failures. - Store attributes in local storage as they are entered - clear local storage when information is saved or modal is closed by the user - (optional): when closing a modal warn users if the values changed since the modal was opened - keep the information in local storage unless modal is closed by the user which addresses application errors, networking hiccups, accidental refresh, computer restarts, etc... - alert the user if the information is restored
1.0
Utilize local storage for an extra level of resiliency to preserve skill/badge/project/subject attributes (ex. description) - Add an extra level of protection when users spend time to create a **new** skill/badge/project/subject attributes followed by the failure to save those changes (due to the networking hiccup, infrastructure instability or a SkillTree bug). Prevent entered information from being lost in the event of unexpected failures. - Store attributes in local storage as they are entered - clear local storage when information is saved or modal is closed by the user - (optional): when closing a modal warn users if the values changed since the modal was opened - keep the information in local storage unless modal is closed by the user which addresses application errors, networking hiccups, accidental refresh, computer restarts, etc... - alert the user if the information is restored
non_process
utilize local storage for an extra level of resiliency to preserve skill badge project subject attributes ex description add an extra level of protection when users spend time to create a new skill badge project subject attributes followed by the failure to save those changes due to the networking hiccup infrastructure instability or a skilltree bug prevent entered information from being lost in the event of unexpected failures store attributes in local storage as they are entered clear local storage when information is saved or modal is closed by the user optional when closing a modal warn users if the values changed since the modal was opened keep the information in local storage unless modal is closed by the user which addresses application errors networking hiccups accidental refresh computer restarts etc alert the user if the information is restored
0
22,908
3,727,389,446
IssuesEvent
2016-03-06 08:05:05
godfather1103/mentohust
https://api.github.com/repos/godfather1103/mentohust
closed
HUST昨天今天一直说版本过低,小女子求救!
auto-migrated Priority-Medium Type-Defect
``` :: 欢迎使用MentoHUST! ** 本机MAC: e8-03-9a-39-37-0b ** 本机IP: 172.20.86.43 ** 子网掩码: 255.255.252.0 ** 默认网关; 172.20.84.1 ** 首选DNS: 202.116.64.107 >> 寻找服务器... ** 认证MAC: 00-1a-a9-17-ff-ff >> 发送用户名... >> 发送密码... >> 认证失败! $$ 系统提示: 锐捷客户端版本过低!A?;牖?u?哠?糛嘾 >> 等待服务器请求或15秒后重启认证。 ** 网关MAC: 00-d0-f8-d5-a8-52 ** 认证MAC: 00-1a-a9-17-ff-ff >> 发送用户名... >> 寻找服务器... >> 发送密码... >> 发送用户名... >> 发送密码... >> 认证失败! $$ 系统提示: 锐捷客户端版本过低!A兖K?漥1?n?糛Om >> 等待服务器请求或15秒后重启认证。 :: 认证已断开! ``` Original issue reported on code.google.com by `newh...@gmail.com` on 15 Jun 2013 at 10:18
1.0
HUST昨天今天一直说版本过低,小女子求救! - ``` :: 欢迎使用MentoHUST! ** 本机MAC: e8-03-9a-39-37-0b ** 本机IP: 172.20.86.43 ** 子网掩码: 255.255.252.0 ** 默认网关; 172.20.84.1 ** 首选DNS: 202.116.64.107 >> 寻找服务器... ** 认证MAC: 00-1a-a9-17-ff-ff >> 发送用户名... >> 发送密码... >> 认证失败! $$ 系统提示: 锐捷客户端版本过低!A?;牖?u?哠?糛嘾 >> 等待服务器请求或15秒后重启认证。 ** 网关MAC: 00-d0-f8-d5-a8-52 ** 认证MAC: 00-1a-a9-17-ff-ff >> 发送用户名... >> 寻找服务器... >> 发送密码... >> 发送用户名... >> 发送密码... >> 认证失败! $$ 系统提示: 锐捷客户端版本过低!A兖K?漥1?n?糛Om >> 等待服务器请求或15秒后重启认证。 :: 认证已断开! ``` Original issue reported on code.google.com by `newh...@gmail.com` on 15 Jun 2013 at 10:18
non_process
hust昨天今天一直说版本过低,小女子求救! 欢迎使用mentohust 本机mac 本机ip 子网掩码 默认网关 首选dns 寻找服务器 认证mac ff ff 发送用户名 发送密码 认证失败 系统提示 锐捷客户端版本过低 a 牖 u 哠 糛嘾 。 网关mac 认证mac ff ff 发送用户名 寻找服务器 发送密码 发送用户名 发送密码 认证失败 系统提示 锐捷客户端版本过低 a兖k n 糛om 。 认证已断开 original issue reported on code google com by newh gmail com on jun at
0
9,087
12,154,523,967
IssuesEvent
2020-04-25 08:45:21
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
opened
`ess-set-working-directory` should not change `default-directory`
process
I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R. It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the file I'm switching from. Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
1.0
`ess-set-working-directory` should not change `default-directory` - I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R. It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the file I'm switching from. Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
process
ess set working directory should not change default directory i ve been confused for years with the behaviour of emacs regarding the active directories of the files i m editing i just realised it s ess set working directory that changes it to the new working directory of r it seems that this behaviour is intended but i m not sure why changing the repl directory should affect the file i m switching from should we change this to preserve the active directory the active directory influences ui commands like m x find file i suspect i m not the only one who is confused by this behaviour
1
7,735
10,854,882,948
IssuesEvent
2019-11-13 17:13:25
codeuniversity/smag-mvp
https://api.github.com/repos/codeuniversity/smag-mvp
opened
Get user data from graphdb using face recognition result
Backend Image Processing
After uploading and analysing the face picture, the data of the user has to be retrieved from graphdb using the encoding.
1.0
Get user data from graphdb using face recognition result - After uploading and analysing the face picture, the data of the user has to be retrieved from graphdb using the encoding.
process
get user data from graphdb using face recognition result after uploading and analysing the face picture the data of the user has to be retrieved from graphdb using the encoding
1
16,155
20,515,161,753
IssuesEvent
2022-03-01 10:57:37
decidim/decidim
https://api.github.com/repos/decidim/decidim
closed
Show statistics and show metrics are always enabled when creating a new process (ignoring admin selection)
type: bug module: participatory processes
**Describe the bug** When I create a new process, "show statistics" and "show metrics" are always enabled regardless I have selected them or not. **To Reproduce** Steps to reproduce the behavior: 1. Go to Processes and Create a New Process 2. Keep unselected "show statistics", "show metrics" or both 3. Create the new process 4. Review process info > "show statistics" and "show metrics" are both selected **Expected behavior** If "show statistics" or "show metrics" are not selected, keep them unselected when creating the process. **Screenshots** https://user-images.githubusercontent.com/72607737/155357502-9c97f784-814d-446e-a730-3e2d9893d391.mov **Extra data (please complete the following information):** - Device: Desktop - Device OS: macOS Catalina - Browser: Chrome & Firefox - Decidim Version: 0.25.2 - Decidim installation: staging.platoniq.net
1.0
Show statistics and show metrics are always enabled when creating a new process (ignoring admin selection) - **Describe the bug** When I create a new process, "show statistics" and "show metrics" are always enabled regardless I have selected them or not. **To Reproduce** Steps to reproduce the behavior: 1. Go to Processes and Create a New Process 2. Keep unselected "show statistics", "show metrics" or both 3. Create the new process 4. Review process info > "show statistics" and "show metrics" are both selected **Expected behavior** If "show statistics" or "show metrics" are not selected, keep them unselected when creating the process. **Screenshots** https://user-images.githubusercontent.com/72607737/155357502-9c97f784-814d-446e-a730-3e2d9893d391.mov **Extra data (please complete the following information):** - Device: Desktop - Device OS: macOS Catalina - Browser: Chrome & Firefox - Decidim Version: 0.25.2 - Decidim installation: staging.platoniq.net
process
show statistics and show metrics are always enabled when creating a new process ignoring admin selection describe the bug when i create a new process show statistics and show metrics are always enabled regardless i have selected them or not to reproduce steps to reproduce the behavior go to processes and create a new process keep unselected show statistics show metrics or both create the new process review process info show statistics and show metrics are both selected expected behavior if show statistics or show metrics are not selected keep them unselected when creating the process screenshots extra data please complete the following information device desktop device os macos catalina browser chrome firefox decidim version decidim installation staging platoniq net
1
11,095
13,937,850,379
IssuesEvent
2020-10-22 14:36:15
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Introduce support for scalar type `BigInt`
engines/data model parser engines/introspection engine engines/migration engine engines/query engine process/candidate team/engines
We decided to introduce the scalar type `BigInt`. So that means that the following schema is not valid anymore: ``` model Blog { id Int @id bigInt Int @mydb.BigInt } ``` Instead users will have to write it like this: ``` model Blog { id Int @id bigInt BigInt } ```
1.0
Introduce support for scalar type `BigInt` - We decided to introduce the scalar type `BigInt`. So that means that the following schema is not valid anymore: ``` model Blog { id Int @id bigInt Int @mydb.BigInt } ``` Instead users will have to write it like this: ``` model Blog { id Int @id bigInt BigInt } ```
process
introduce support for scalar type bigint we decided to introduce the scalar type bigint so that means that the following schema is not valid anymore model blog id int id bigint int mydb bigint instead users will have to write it like this model blog id int id bigint bigint
1
2,216
11,592,816,185
IssuesEvent
2020-02-24 12:20:08
big-neon/bn-web
https://api.github.com/repos/big-neon/bn-web
opened
Automation: Big Neon: Test: Allow Past Events to Remain On Site: Search Using URL
Automation
**Pre-conditions:** 1. User should have admin access to Big Neon 2. User should be logged into Big Neon 3. User should have an event that has past **Steps:** 1. Add the URL to view the event that has past 2. View event page loads successfully 3. Verify the button "Purchase Tickets" is now displayed as "This Event Is Now Over" with a tear emoji 4. Try to select the above button 5. User should be unable to select the button test pad link: https://big-neon.ontestpad.com/script/194#11//
1.0
Automation: Big Neon: Test: Allow Past Events to Remain On Site: Search Using URL - **Pre-conditions:** 1. User should have admin access to Big Neon 2. User should be logged into Big Neon 3. User should have an event that has past **Steps:** 1. Add the URL to view the event that has past 2. View event page loads successfully 3. Verify the button "Purchase Tickets" is now displayed as "This Event Is Now Over" with a tear emoji 4. Try to select the above button 5. User should be unable to select the button test pad link: https://big-neon.ontestpad.com/script/194#11//
non_process
automation big neon test allow past events to remain on site search using url pre conditions user should have admin access to big neon user should be logged into big neon user should have an event that has past steps add the url to view the event that has past view event page loads successfully verify the button purchase tickets is now displayed as this event is now over with a tear emoji try to select the above button user should be unable to select the button test pad link
0
303,863
9,311,343,138
IssuesEvent
2019-03-25 21:07:08
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
Not able to find merger for match combination when conflating Bandug data
Category: Algorithms Priority: Medium Status: Defined Type: Bug
Must run with network conflation for this (see command below). This is the follow up task to #2059 which put a bandaid on the problem. It is possible that the fixes for #2059 are valid by themselves, but that needs to be proven as part of this task. Input data: ref: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/MapEditBandug.osm sec: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/OSMmap.osm If you re-enable the exception that used to be thrown in MergerFactory, you will see that a NetworkMatch and a ScriptMatch are grouped together to be merged, which I don't think makes sense. Since no merger can merge those two, the exception ends up being thrown. If this is proven to be a bug and gets fixed, you can also remove the extra error checking that was added to ScriptMerger::_applyMergePair as part of #2059. `hoot crop-map MapEditBandug.osm MapEditBandug-cropped.osm "107.604,-6.922,107.611,-6.915"` `hoot crop-map OSMmap.osm OSMmap-cropped.osm "107.604,-6.922,107.611,-6.915"` `hoot conflate -D uuid.helper.repeatable=true -D writer.include.debug.tags=true -D match.creators="hoot::BuildingMatchCreator;hoot::ScriptMatchCreator,PoiGeneric.js;hoot::NetworkMatchCreator;hoot::ScriptMatchCreator,LinearWaterway.js;hoot::PoiPolygonMatchCreator;hoot::ScriptMatchCreator,Area.js" -D merger.creators="hoot::BuildingMergerCreator;hoot::ScriptMergerCreator;hoot::NetworkMergerCreator;hoot::PoiPolygonMergerCreator;hoot::ScriptMergerCreator" -D conflate.match.highway.classifier="hoot::HighwayExpertClassifier" -D way.subline.matcher="hoot::MaximalSublineMatcher" -D rubber.sheet.minimum.ties=4 -D rubber.sheet.ref=true MapEditBandug-cropped.osm OSMmap-cropped.osm out.osm`
1.0
Not able to find merger for match combination when conflating Bandug data - Must run with network conflation for this (see command below). This is the follow up task to #2059 which put a bandaid on the problem. It is possible that the fixes for #2059 are valid by themselves, but that needs to be proven as part of this task. Input data: ref: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/MapEditBandug.osm sec: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/OSMmap.osm If you re-enable the exception that used to be thrown in MergerFactory, you will see that a NetworkMatch and a ScriptMatch are grouped together to be merged, which I don't think makes sense. Since no merger can merge those two, the exception ends up being thrown. If this is proven to be a bug and gets fixed, you can also remove the extra error checking that was added to ScriptMerger::_applyMergePair as part of #2059. `hoot crop-map MapEditBandug.osm MapEditBandug-cropped.osm "107.604,-6.922,107.611,-6.915"` `hoot crop-map OSMmap.osm OSMmap-cropped.osm "107.604,-6.922,107.611,-6.915"` `hoot conflate -D uuid.helper.repeatable=true -D writer.include.debug.tags=true -D match.creators="hoot::BuildingMatchCreator;hoot::ScriptMatchCreator,PoiGeneric.js;hoot::NetworkMatchCreator;hoot::ScriptMatchCreator,LinearWaterway.js;hoot::PoiPolygonMatchCreator;hoot::ScriptMatchCreator,Area.js" -D merger.creators="hoot::BuildingMergerCreator;hoot::ScriptMergerCreator;hoot::NetworkMergerCreator;hoot::PoiPolygonMergerCreator;hoot::ScriptMergerCreator" -D conflate.match.highway.classifier="hoot::HighwayExpertClassifier" -D way.subline.matcher="hoot::MaximalSublineMatcher" -D rubber.sheet.minimum.ties=4 -D rubber.sheet.ref=true MapEditBandug-cropped.osm OSMmap-cropped.osm out.osm`
non_process
not able to find merger for match combination when conflating bandug data must run with network conflation for this see command below this is the follow up task to which put a bandaid on the problem it is possible that the fixes for are valid by themselves but that needs to be proven as part of this task input data ref sec if you re enable the exception that used to be thrown in mergerfactory you will see that a networkmatch and a scriptmatch are grouped together to be merged which i don t think makes sense since no merger can merge those two the exception ends up being thrown if this is proven to be a bug and gets fixed you can also remove the extra error checking that was added to scriptmerger applymergepair as part of hoot crop map mapeditbandug osm mapeditbandug cropped osm hoot crop map osmmap osm osmmap cropped osm hoot conflate d uuid helper repeatable true d writer include debug tags true d match creators hoot buildingmatchcreator hoot scriptmatchcreator poigeneric js hoot networkmatchcreator hoot scriptmatchcreator linearwaterway js hoot poipolygonmatchcreator hoot scriptmatchcreator area js d merger creators hoot buildingmergercreator hoot scriptmergercreator hoot networkmergercreator hoot poipolygonmergercreator hoot scriptmergercreator d conflate match highway classifier hoot highwayexpertclassifier d way subline matcher hoot maximalsublinematcher d rubber sheet minimum ties d rubber sheet ref true mapeditbandug cropped osm osmmap cropped osm out osm
0
5,810
8,644,718,444
IssuesEvent
2018-11-26 04:38:01
gfrebello/qs-trip-planning-procedure
https://api.github.com/repos/gfrebello/qs-trip-planning-procedure
closed
Realize Tests and Update Test Document
Priority:High Process:Run Test Case
Tests need to be realized to make sure the requirements are being met, and if not, either the project needs to be corrected, or the requirements need to be changed.
1.0
Realize Tests and Update Test Document - Tests need to be realized to make sure the requirements are being met, and if not, either the project needs to be corrected, or the requirements need to be changed.
process
realize tests and update test document tests need to be realized to make sure the requirements are being met and if not either the project needs to be corrected or the requirements need to be changed
1
234,161
19,098,621,353
IssuesEvent
2021-11-29 19:34:20
microsoft/vscode-python
https://api.github.com/repos/microsoft/vscode-python
opened
Test Pylance with a virtual workspace
testplan-item
Refs: #17253 - [ ] anyOS - [ ] anyOS - [ ] anyOS Complexity: 2 Authors: @karrtikr --- ### Requirements 1. Install python extension from here: https://pvsc.blob.core.windows.net/extension-builds/ms-python-insiders.vsix ### Verification 1. Open a GitHub repository containing Python code, and make sure it is trusted 2. Make sure that you have the `python.languageServer` setting set to `Default` or `Pylance` 3. Activate the Python extension 4. Verify that the `Starting Pylance language server.` message appears in the Python output channel 5. Make sure you get hover info, same-file completion and completion for built-ins like `print`
1.0
Test Pylance with a virtual workspace - Refs: #17253 - [ ] anyOS - [ ] anyOS - [ ] anyOS Complexity: 2 Authors: @karrtikr --- ### Requirements 1. Install python extension from here: https://pvsc.blob.core.windows.net/extension-builds/ms-python-insiders.vsix ### Verification 1. Open a GitHub repository containing Python code, and make sure it is trusted 2. Make sure that you have the `python.languageServer` setting set to `Default` or `Pylance` 3. Activate the Python extension 4. Verify that the `Starting Pylance language server.` message appears in the Python output channel 5. Make sure you get hover info, same-file completion and completion for built-ins like `print`
non_process
test pylance with a virtual workspace refs anyos anyos anyos complexity authors karrtikr requirements install python extension from here verification open a github repository containing python code and make sure it is trusted make sure that you have the python languageserver setting set to default or pylance activate the python extension verify that the starting pylance language server message appears in the python output channel make sure you get hover info same file completion and completion for built ins like print
0
268,479
23,373,633,030
IssuesEvent
2022-08-10 22:51:36
kubernetes/test-infra
https://api.github.com/repos/kubernetes/test-infra
closed
Job names have wrong suffix in new branch
kind/bug sig/testing sig/release
**What happened**: When generating the new 1.25 branch jobs, the names of the jobs were wrongly generated, causing a whole bunch of presubmits to fail when verifying the config files. The job names had the release version appended (`-1.25`) **What you expected to happen**: Job names should not have had the suffixes added **How to reproduce it (as minimally and precisely as possible)**: Ran `make -C releng prepare-release-branch` **Please provide links to example occurrences, if any**: https://github.com/kubernetes/test-infra/pull/27096/commits/8058610ea3dd366a2153f5d8a8bf0adfc81d39d9 **Anything else we need to know?**:
1.0
Job names have wrong suffix in new branch - **What happened**: When generating the new 1.25 branch jobs, the names of the jobs were wrongly generated, causing a whole bunch of presubmits to fail when verifying the config files. The job names had the release version appended (`-1.25`) **What you expected to happen**: Job names should not have had the suffixes added **How to reproduce it (as minimally and precisely as possible)**: Ran `make -C releng prepare-release-branch` **Please provide links to example occurrences, if any**: https://github.com/kubernetes/test-infra/pull/27096/commits/8058610ea3dd366a2153f5d8a8bf0adfc81d39d9 **Anything else we need to know?**:
non_process
job names have wrong suffix in new branch what happened when generating the new branch jobs the names of the jobs were wrongly generated causing a whole bunch of presubmits to fail when verifying the config files the job names had the release version appended what you expected to happen job names should not have had the suffixes added how to reproduce it as minimally and precisely as possible ran make c releng prepare release branch please provide links to example occurrences if any anything else we need to know
0
16,279
20,884,554,008
IssuesEvent
2022-03-23 02:34:50
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add All Over The Girl
suggested title in process
Please add as much of the following info as you can: Title: All Over The Girl Type (film/tv show): Film Film or show in which it appears: Marriage Story Is the parent film/show streaming anywhere? [Netflix](https://www.netflix.com/title/80223779) About when in the parent film/show does it appear? 2m50s Actual footage of the film/show can be seen (yes/no)? Yes
1.0
Add All Over The Girl - Please add as much of the following info as you can: Title: All Over The Girl Type (film/tv show): Film Film or show in which it appears: Marriage Story Is the parent film/show streaming anywhere? [Netflix](https://www.netflix.com/title/80223779) About when in the parent film/show does it appear? 2m50s Actual footage of the film/show can be seen (yes/no)? Yes
process
add all over the girl please add as much of the following info as you can title all over the girl type film tv show film film or show in which it appears marriage story is the parent film show streaming anywhere about when in the parent film show does it appear actual footage of the film show can be seen yes no yes
1
661,559
22,060,677,611
IssuesEvent
2022-05-30 17:21:25
SimpleGeek/squid-config-ui
https://api.github.com/repos/SimpleGeek/squid-config-ui
opened
Bypass menu not correctly displaying time after key sequence
bug Priority: Medium
To reproduce: Enter a valid time, then hit enter again. You should get the error alert that the time was invalid, and then the bypass list should say there are no active bypasses, despite a bypass being active for the time assigned in the first step.
1.0
Bypass menu not correctly displaying time after key sequence - To reproduce: Enter a valid time, then hit enter again. You should get the error alert that the time was invalid, and then the bypass list should say there are no active bypasses, despite a bypass being active for the time assigned in the first step.
non_process
bypass menu not correctly displaying time after key sequence to reproduce enter a valid time then hit enter again you should get the error alert that the time was invalid and then the bypass list should say there are no active bypasses despite a bypass being active for the time assigned in the first step
0
19,794
26,178,260,419
IssuesEvent
2023-01-02 12:31:29
mdsreq-fga-unb/2022.2-Receitalista
https://api.github.com/repos/mdsreq-fga-unb/2022.2-Receitalista
closed
Colocar as atividades de cada evento Scrum
processo visao
No tópico 3.2 do documento da Visão, deve-se separar os eventos do Scrum para colocar as atividades de cada evento. **Como está:** |Atividade|Método|Ferramenta|Entrega| |:-----:|:------:|:------:|:------:| |Planejamento do Sprint (sprint planning)|Reunião|Discord|Backlog do Sprint| |Revisão do Sprint (sprint review)|Reunião|Discord|Nova release| |Reunião diária (daily)|Conversa entre os integrantes da equipe|Telegram|Feedback do Sprint| |Retrospectiva do Sprint (sprint retrospecive)|Reunião|Discord|Feedback do Sprint| **Como deve ser:** ### 3.n - Planejamento do Sprint (sprint planning) |Atividade|Método|Ferramenta|Entrega| |:-----:|:------:|:------:|:------:| |Atividade 1|etc|etc|etc| |Atividade 2|etc|etc|etc| ### 3.n - Revisão do Sprint (sprint review) |Atividade|Método|Ferramenta|Entrega| |:-----:|:------:|:------:|:------:| |Atividade 1|etc|etc|etc| |Atividade 2|etc|etc|etc|
1.0
Colocar as atividades de cada evento Scrum - No tópico 3.2 do documento da Visão, deve-se separar os eventos do Scrum para colocar as atividades de cada evento. **Como está:** |Atividade|Método|Ferramenta|Entrega| |:-----:|:------:|:------:|:------:| |Planejamento do Sprint (sprint planning)|Reunião|Discord|Backlog do Sprint| |Revisão do Sprint (sprint review)|Reunião|Discord|Nova release| |Reunião diária (daily)|Conversa entre os integrantes da equipe|Telegram|Feedback do Sprint| |Retrospectiva do Sprint (sprint retrospecive)|Reunião|Discord|Feedback do Sprint| **Como deve ser:** ### 3.n - Planejamento do Sprint (sprint planning) |Atividade|Método|Ferramenta|Entrega| |:-----:|:------:|:------:|:------:| |Atividade 1|etc|etc|etc| |Atividade 2|etc|etc|etc| ### 3.n - Revisão do Sprint (sprint review) |Atividade|Método|Ferramenta|Entrega| |:-----:|:------:|:------:|:------:| |Atividade 1|etc|etc|etc| |Atividade 2|etc|etc|etc|
process
colocar as atividades de cada evento scrum no tópico do documento da visão deve se separar os eventos do scrum para colocar as atividades de cada evento como está atividade método ferramenta entrega planejamento do sprint sprint planning reunião discord backlog do sprint revisão do sprint sprint review reunião discord nova release reunião diária daily conversa entre os integrantes da equipe telegram feedback do sprint retrospectiva do sprint sprint retrospecive reunião discord feedback do sprint como deve ser n planejamento do sprint sprint planning atividade método ferramenta entrega atividade etc etc etc atividade etc etc etc n revisão do sprint sprint review atividade método ferramenta entrega atividade etc etc etc atividade etc etc etc
1
1,112
9,488,823,580
IssuesEvent
2019-04-22 20:38:21
mozilla-mobile/reference-browser
https://api.github.com/repos/mozilla-mobile/reference-browser
closed
Dep-sign debug builds on-push on the master branch
🤖 automation
We will have performance tests running on `reference-browser` in the future. One of the requirements is to have builds signed with the dep key. This ticket requires that, on every commit to the master branch, all the debug variants should be dep-signed.
1.0
Dep-sign debug builds on-push on the master branch - We will have performance tests running on `reference-browser` in the future. One of the requirements is to have builds signed with the dep key. This ticket requires that, on every commit to the master branch, all the debug variants should be dep-signed.
non_process
dep sign debug builds on push on the master branch we will have performance tests running on reference browser in the future one of the requirements is to have builds signed with the dep key this ticket requires that on every commit to the master branch all the debug variants should be dep signed
0
3,774
4,043,821,506
IssuesEvent
2016-05-21 00:13:57
google/cadvisor
https://api.github.com/repos/google/cadvisor
closed
[scalability] failed to collect filesystem stats - failed to exec du
area/performance
Full error: ``` E0520 20:38:01.903710 16716 fsHandler.go:106] failed to collect filesystem stats - failed to exec du - fork/exec /usr/bin/nice: cannot allocate memory ``` To reproduce: ``` $ sudo ./cadvisor --logtostderr & $ for i in `seq 1 500`; do docker run -d kubernetes/pause; done ``` I suggest we limit the number of parallel execs to a reasonable number.
True
[scalability] failed to collect filesystem stats - failed to exec du - Full error: ``` E0520 20:38:01.903710 16716 fsHandler.go:106] failed to collect filesystem stats - failed to exec du - fork/exec /usr/bin/nice: cannot allocate memory ``` To reproduce: ``` $ sudo ./cadvisor --logtostderr & $ for i in `seq 1 500`; do docker run -d kubernetes/pause; done ``` I suggest we limit the number of parallel execs to a reasonable number.
non_process
failed to collect filesystem stats failed to exec du full error fshandler go failed to collect filesystem stats failed to exec du fork exec usr bin nice cannot allocate memory to reproduce sudo cadvisor logtostderr for i in seq do docker run d kubernetes pause done i suggest we limit the number of parallel execs to a reasonable number
0
10,306
13,155,329,080
IssuesEvent
2020-08-10 08:41:19
didi/mpx
https://api.github.com/repos/didi/mpx
closed
[Bug report] style scoped page选择器能否过滤
processing
**问题描述** 使用autoScopeRules时如果页面文件中包含页面,支付宝下页面中含有page选择器会也会加上scopedId,导致page选择器无效,scoped这里可以针对page选择器进行过滤吗? **环境信息描述** Mac mpx版本 "@mpxjs/core": "^2.5.33" "@mpxjs/webpack-plugin": "^2.5.33" 支付宝IDE 1.13.4 **最简复现demo** ```javascript // mpx.webpack.conf.js { autoScopeRules: { include: [resolve('src')], exclude: [ resolve('src/app.mpx'), resolve('src/assets'), resolve('src/components/@vant'), ], }, } ``` ```css // src/pages/xxx.mpx page { background-color: #f8f8f8 } // 编译后 page.m9ed7a486 { background-color: #f8f8f8; } ```
1.0
[Bug report] style scoped page选择器能否过滤 - **问题描述** 使用autoScopeRules时如果页面文件中包含页面,支付宝下页面中含有page选择器会也会加上scopedId,导致page选择器无效,scoped这里可以针对page选择器进行过滤吗? **环境信息描述** Mac mpx版本 "@mpxjs/core": "^2.5.33" "@mpxjs/webpack-plugin": "^2.5.33" 支付宝IDE 1.13.4 **最简复现demo** ```javascript // mpx.webpack.conf.js { autoScopeRules: { include: [resolve('src')], exclude: [ resolve('src/app.mpx'), resolve('src/assets'), resolve('src/components/@vant'), ], }, } ``` ```css // src/pages/xxx.mpx page { background-color: #f8f8f8 } // 编译后 page.m9ed7a486 { background-color: #f8f8f8; } ```
process
style scoped page选择器能否过滤 问题描述 使用autoscoperules时如果页面文件中包含页面,支付宝下页面中含有page选择器会也会加上scopedid,导致page选择器无效,scoped这里可以针对page选择器进行过滤吗? 环境信息描述 mac mpx版本 mpxjs core mpxjs webpack plugin 支付宝ide 最简复现demo javascript mpx webpack conf js autoscoperules include exclude resolve src app mpx resolve src assets resolve src components vant css src pages xxx mpx page background color 编译后 page background color
1
14,376
17,398,554,278
IssuesEvent
2021-08-02 16:17:15
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
Release 0.3.0
process + tools
This is a coordinating issue for the release. Issues for this release are listed here: https://github.com/pystatgen/sgkit/milestone/1 The artifacts for the previous release (#495) were built manually. For this release I would like to build and deploy artifacts using GH actions, see #606 and #607.
1.0
Release 0.3.0 - This is a coordinating issue for the release. Issues for this release are listed here: https://github.com/pystatgen/sgkit/milestone/1 The artifacts for the previous release (#495) were built manually. For this release I would like to build and deploy artifacts using GH actions, see #606 and #607.
process
release this is a coordinating issue for the release issues for this release are listed here the artifacts for the previous release were built manually for this release i would like to build and deploy artifacts using gh actions see and
1
365,346
10,780,931,830
IssuesEvent
2019-11-04 13:59:53
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
closed
/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md
finished lang/zh priority/P0 sync/new version/1.16
Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md)
1.0
/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md)
non_process
docs reference setup tools kubeadm generated kubeadm reset phase preflight md source file
0
14,369
17,394,251,102
IssuesEvent
2021-08-02 11:26:29
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Source Control Sync Job failed with "The property 'data' cannot be found on this object" message
Pri2 automation/svc cxp process-automation/subsvc product-issue triaged
We have a Source Control Sync setup with GitHub. Until recently it was working properly. Even though the sync job is shown as completed, there is an error in the logs: ``` System.Management.Automation.PropertyNotFoundException: The property 'data' cannot be found on this object. Verify that the property exists. at System.Management.Automation.ExceptionHandlingOps.CheckActionPreference(FunctionContext funcContext, Exception exception) at System.Management.Automation.Interpreter.ActionCallInstruction`2.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.Interpreter.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.LightLambda.RunVoid1[T0](T0 arg0) at System.Management.Automation.PSScriptCmdlet.RunClause(Action`1 clause, Object dollarUnderbar, Object inputToProcess) at System.Management.Automation.PSScriptCmdlet.DoEndProcessing() at System.Management.Automation.CommandProcessorBase.Complete() ``` The sync of course is not being done. Thank you! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Source Control Sync Job failed with "The property 'data' cannot be found on this object" message - We have a Source Control Sync setup with GitHub. Until recently it was working properly. Even though the sync job is shown as completed, there is an error in the logs: ``` System.Management.Automation.PropertyNotFoundException: The property 'data' cannot be found on this object. Verify that the property exists. at System.Management.Automation.ExceptionHandlingOps.CheckActionPreference(FunctionContext funcContext, Exception exception) at System.Management.Automation.Interpreter.ActionCallInstruction`2.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.Interpreter.Run(InterpretedFrame frame) at System.Management.Automation.Interpreter.LightLambda.RunVoid1[T0](T0 arg0) at System.Management.Automation.PSScriptCmdlet.RunClause(Action`1 clause, Object dollarUnderbar, Object inputToProcess) at System.Management.Automation.PSScriptCmdlet.DoEndProcessing() at System.Management.Automation.CommandProcessorBase.Complete() ``` The sync of course is not being done. Thank you! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
source control sync job failed with the property data cannot be found on this object message we have a source control sync setup with github until recently it was working properly even though the sync job is shown as completed there is an error in the logs system management automation propertynotfoundexception the property data cannot be found on this object verify that the property exists at system management automation exceptionhandlingops checkactionpreference functioncontext funccontext exception exception at system management automation interpreter actioncallinstruction run interpretedframe frame at system management automation interpreter entertrycatchfinallyinstruction run interpretedframe frame at system management automation interpreter entertrycatchfinallyinstruction run interpretedframe frame at system management automation interpreter interpreter run interpretedframe frame at system management automation interpreter lightlambda at system management automation psscriptcmdlet runclause action clause object dollarunderbar object inputtoprocess at system management automation psscriptcmdlet doendprocessing at system management automation commandprocessorbase complete the sync of course is not being done thank you document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
20,005
26,479,558,733
IssuesEvent
2023-01-17 13:46:25
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
bazel-1.0.0: Depends: python but it is not installable
P2 type: support / not a bug (process) team-OSS
### Description of the bug: I'm trying the install bazel 1.0.0 (a C++ package depends on this version specifically). When I attempt to install it I get the following message: `$ sudo apt install bazel-1.0.0 Reading package lists... Done Building dependency tree... Done Reading state information... Done Some packages could not be installed. This may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of Incoming. The following information may help to resolve the situation: The following packages have unmet dependencies: bazel-1.0.0 : Depends: python but it is not installable E: Unable to correct problems, you have held broken packages.` I have python 3.10.6 installed on my system: `python Python 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> exit()` All other dependences are installed. My system is Ubuntu command line installed through WSL2: `$ lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 22.04.1 LTS Release: 22.04 Codename: jammy` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Attempt to install Bazel 1.0.0 on Ubuntu 22.04 through WSL2 on Windows 11. ### Which operating system are you running Bazel on? Ubuntu 22.04 - jammy ### What is the output of `bazel info release`? NA ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. NA ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? ```text NA ``` ### Have you found anything relevant by searching the web? There was a closed submission on this site. No resolution was listed. ### Any other information, logs, or outputs that you want to share? _No response_
1.0
bazel-1.0.0: Depends: python but it is not installable - ### Description of the bug: I'm trying the install bazel 1.0.0 (a C++ package depends on this version specifically). When I attempt to install it I get the following message: `$ sudo apt install bazel-1.0.0 Reading package lists... Done Building dependency tree... Done Reading state information... Done Some packages could not be installed. This may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of Incoming. The following information may help to resolve the situation: The following packages have unmet dependencies: bazel-1.0.0 : Depends: python but it is not installable E: Unable to correct problems, you have held broken packages.` I have python 3.10.6 installed on my system: `python Python 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> exit()` All other dependences are installed. My system is Ubuntu command line installed through WSL2: `$ lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 22.04.1 LTS Release: 22.04 Codename: jammy` ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Attempt to install Bazel 1.0.0 on Ubuntu 22.04 through WSL2 on Windows 11. ### Which operating system are you running Bazel on? Ubuntu 22.04 - jammy ### What is the output of `bazel info release`? NA ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. NA ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? ```text NA ``` ### Have you found anything relevant by searching the web? There was a closed submission on this site. No resolution was listed. ### Any other information, logs, or outputs that you want to share? _No response_
process
bazel depends python but it is not installable description of the bug i m trying the install bazel a c package depends on this version specifically when i attempt to install it i get the following message sudo apt install bazel reading package lists done building dependency tree done reading state information done some packages could not be installed this may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of incoming the following information may help to resolve the situation the following packages have unmet dependencies bazel depends python but it is not installable e unable to correct problems you have held broken packages i have python installed on my system python python main nov on linux type help copyright credits or license for more information exit all other dependences are installed my system is ubuntu command line installed through lsb release a no lsb modules are available distributor id ubuntu description ubuntu lts release codename jammy what s the simplest easiest way to reproduce this bug please provide a minimal example if possible attempt to install bazel on ubuntu through on windows which operating system are you running bazel on ubuntu jammy what is the output of bazel info release na if bazel info release returns development version or non git tell us how you built bazel na what s the output of git remote get url origin git rev parse master git rev parse head text na have you found anything relevant by searching the web there was a closed submission on this site no resolution was listed any other information logs or outputs that you want to share no response
1
345,713
24,872,356,160
IssuesEvent
2022-10-27 16:07:45
Telefonica/mistica-design
https://api.github.com/repos/Telefonica/mistica-design
opened
toggleButton component
component: buttons library: mobile library: desktop category: documentation platform: ios platform: android platform: web category: materials new
- [ ] Specs - [ ] Design documentation - [ ] Add to mobile library - [ ] Add to desktop library - [ ] Add to UI Kit - [ ] Android ticket - [ ] iOS ticket - [ ] Web ticket
1.0
toggleButton component - - [ ] Specs - [ ] Design documentation - [ ] Add to mobile library - [ ] Add to desktop library - [ ] Add to UI Kit - [ ] Android ticket - [ ] iOS ticket - [ ] Web ticket
non_process
togglebutton component specs design documentation add to mobile library add to desktop library add to ui kit android ticket ios ticket web ticket
0
117,774
25,194,327,030
IssuesEvent
2022-11-12 10:00:10
rocky/python-decompile3
https://api.github.com/repos/rocky/python-decompile3
closed
Parse error at or near `DUP_TOP' (python 3.8)
invalid bytecode suspicious activity
When I decompiled the pyc, got the following error messages. OS : win 10 Python 3.8.6 uncompyle6 : 3.7.4 decompyle3 : 3.3.2 pydisasm version 5.0.4 Python bytecode 3.8 (3413) Source code size mod 2**32 ``` Instruction context: L. 5 52 LOAD_NAME list 54 LOAD_GENEXPR '<code_object <genexpr>>' 56 LOAD_STR '<genexpr>' 58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 60 LOAD_GENEXPR '<code_object <genexpr>>' 62 LOAD_STR '<genexpr>' 64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 66 LOAD_NAME f_in 68 GET_ITER 70 CALL_FUNCTION_1 1 '' 72 GET_ITER 74 CALL_FUNCTION_1 1 '' 76 CALL_FUNCTION_1 1 '' 78 STORE_NAME requirements 80 POP_BLOCK 82 LOAD_CONST None -> 84 DUP_TOP 86 DUP_TOP 88 CALL_FUNCTION_3 3 '' 90 POP_TOP 92 JUMP_FORWARD 110 'to 110' 94_0 COME_FROM_WITH 48 '48' 94 <49> 96 POP_JUMP_IF_TRUE 100 'to 100' 98 <48> 100_0 COME_FROM 96 '96' 100 POP_TOP 102 POP_TOP 104 POP_TOP 106 POP_EXCEPT 108 POP_TOP 110_0 COME_FROM 92 '92' # file start.pyc # --- This code section failed: --- L. 2 0 LOAD_CONST 0 2 LOAD_CONST None 4 IMPORT_NAME re 6 STORE_NAME re 8 LOAD_CONST 0 10 LOAD_CONST None 12 IMPORT_NAME os 14 STORE_NAME os 16 LOAD_CONST 0 18 LOAD_CONST None 20 IMPORT_NAME sys 22 STORE_NAME sys 24 LOAD_CONST 0 26 LOAD_CONST None 28 IMPORT_NAME pkg_resources 30 STORE_NAME pkg_resources 32 LOAD_CONST 0 34 LOAD_CONST None 36 IMPORT_NAME random 38 STORE_NAME random L. 3 40 SETUP_FINALLY 228 'to 228' L. 4 42 LOAD_NAME open 44 LOAD_STR 'requirements.txt' 46 CALL_FUNCTION_1 1 '' 48 SETUP_WITH 94 'to 94' 50 STORE_NAME f_in L. 5 52 LOAD_NAME list 54 LOAD_GENEXPR '<code_object <genexpr>>' 56 LOAD_STR '<genexpr>' 58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 60 LOAD_GENEXPR '<code_object <genexpr>>' 62 LOAD_STR '<genexpr>' 64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 66 LOAD_NAME f_in 68 GET_ITER 70 CALL_FUNCTION_1 1 '' 72 GET_ITER 74 CALL_FUNCTION_1 1 '' 76 CALL_FUNCTION_1 1 '' 78 STORE_NAME requirements 80 POP_BLOCK 82 LOAD_CONST None 84 DUP_TOP 86 DUP_TOP 88 CALL_FUNCTION_3 3 '' 90 POP_TOP 92 JUMP_FORWARD 110 'to 110' 94_0 COME_FROM_WITH 48 '48' 94 <49> 96 POP_JUMP_IF_TRUE 100 'to 100' 98 <48> 100_0 COME_FROM 96 '96' 100 POP_TOP 102 POP_TOP 104 POP_TOP 106 POP_EXCEPT 108 POP_TOP 110_0 COME_FROM 92 '92' L. 6 110 LOAD_NAME requirements 112 GET_ITER 114_0 COME_FROM 222 '222' 114_1 COME_FROM 210 '210' 114_2 COME_FROM 132 '132' 114 FOR_ITER 224 'to 224' 116 STORE_NAME require L. 7 118 SETUP_FINALLY 134 'to 134' L. 8 120 LOAD_NAME pkg_resources 122 LOAD_METHOD require 124 LOAD_NAME require 126 CALL_METHOD_1 1 '' 128 POP_TOP 130 POP_BLOCK 132 JUMP_BACK 114 'to 114' 134_0 COME_FROM_FINALLY 118 '118' L. 9 134 DUP_TOP 136 LOAD_NAME Exception 138 <121> 220 '' 140 POP_TOP 142 STORE_NAME e 144 POP_TOP 146 SETUP_FINALLY 212 'to 212' L. 10 148 LOAD_NAME re 150 LOAD_METHOD search 152 LOAD_STR "\\'(.*?)\\'" 154 LOAD_NAME str 156 LOAD_NAME e 158 CALL_FUNCTION_1 1 '' 160 CALL_METHOD_2 2 '' 162 LOAD_METHOD group 164 LOAD_CONST 1 166 CALL_METHOD_1 1 '' 168 STORE_NAME a L. 11 170 LOAD_NAME print 172 LOAD_STR 'Installing ' 174 LOAD_NAME a 176 BINARY_ADD 178 LOAD_STR ' ...' 180 BINARY_ADD 182 CALL_FUNCTION_1 1 '' 184 POP_TOP L. 12 186 LOAD_NAME os 188 LOAD_METHOD system 190 LOAD_STR 'pip install ' 192 LOAD_NAME a 194 BINARY_ADD 196 CALL_METHOD_1 1 '' 198 POP_TOP 200 POP_BLOCK 202 POP_EXCEPT 204 LOAD_CONST None 206 STORE_NAME e 208 DELETE_NAME e 210 JUMP_BACK 114 'to 114' 212_0 COME_FROM_FINALLY 146 '146' 212 LOAD_CONST None 214 STORE_NAME e 216 DELETE_NAME e 218 <48> 220 <48> 222 JUMP_BACK 114 'to 114' 224_0 COME_FROM 114 '114' 224 POP_BLOCK 226 JUMP_FORWARD 280 'to 280' 228_0 COME_FROM_FINALLY 40 '40' L. 13 228 DUP_TOP 230 LOAD_NAME Exception 232_234 <121> 278 '' 236 POP_TOP 238 STORE_NAME e 240 POP_TOP 242 SETUP_FINALLY 270 'to 270' L. 14 244 LOAD_NAME print 246 LOAD_STR 'installation error: ' 248 LOAD_NAME e 250 FORMAT_VALUE 0 '' 252 BUILD_STRING_2 2 254 CALL_FUNCTION_1 1 '' 256 POP_TOP 258 POP_BLOCK 260 POP_EXCEPT 262 LOAD_CONST None 264 STORE_NAME e 266 DELETE_NAME e 268 JUMP_FORWARD 280 'to 280' 270_0 COME_FROM_FINALLY 242 '242' 270 LOAD_CONST None 272 STORE_NAME e 274 DELETE_NAME e 276 <48> 278 <48> 280_0 COME_FROM 268 '268' 280_1 COME_FROM 226 '226' L. 16 280 LOAD_CONST 0 282 LOAD_CONST None 284 IMPORT_NAME asyncio 286 STORE_NAME asyncio 288 LOAD_CONST 0 290 LOAD_CONST None 292 IMPORT_NAME logging 294 STORE_NAME logging 296 LOAD_CONST 0 298 LOAD_CONST None 300 IMPORT_NAME requests 302 STORE_NAME requests 304 LOAD_CONST 0 306 LOAD_CONST None 308 IMPORT_NAME sqlite3 310 STORE_NAME sqlite3 312 LOAD_CONST 0 314 LOAD_CONST None 316 IMPORT_NAME base64 318 STORE_NAME base64 L. 17 320 LOAD_CONST 0 322 LOAD_CONST ('sleep', 'strptime') 324 IMPORT_NAME time 326 IMPORT_FROM sleep 328 STORE_NAME sleep 330 IMPORT_FROM strptime 332 STORE_NAME strptime 334 POP_TOP L. 18 336 LOAD_CONST 0 338 LOAD_CONST ('KeyboardButtonUrl', 'MessageMediaContact', 'ReplyInlineMarkup', 'UpdateShortMessage') 340 IMPORT_NAME_ATTR telethon.tl.types 342 IMPORT_FROM KeyboardButtonUrl 344 STORE_NAME KeyboardButtonUrl 346 IMPORT_FROM MessageMediaContact 348 STORE_NAME MessageMediaContact 350 IMPORT_FROM ReplyInlineMarkup 352 STORE_NAME ReplyInlineMarkup 354 IMPORT_FROM UpdateShortMessage 356 STORE_NAME UpdateShortMessage 358 POP_TOP L. 19 360 LOAD_CONST 0 362 LOAD_CONST ('TelegramClient', 'client', 'errors', 'events', 'functions', 'connection') 364 IMPORT_NAME telethon 366 IMPORT_FROM TelegramClient 368 STORE_NAME TelegramClient 370 IMPORT_FROM client 372 STORE_GLOBAL client 374 IMPORT_FROM errors 376 STORE_NAME errors 378 IMPORT_FROM events 380 STORE_NAME events 382 IMPORT_FROM functions 384 STORE_NAME functions 386 IMPORT_FROM connection 388 STORE_NAME connection 390 POP_TOP L. 20 392 LOAD_CONST 0 394 LOAD_CONST ('JoinChannelRequest',) 396 IMPORT_NAME_ATTR telethon.tl.functions.channels 398 IMPORT_FROM JoinChannelRequest 400 STORE_NAME JoinChannelRequest 402 POP_TOP L. 21 404 LOAD_CONST 0 406 LOAD_CONST ('GetBotCallbackAnswerRequest', 'GetHistoryRequest', 'ImportChatInviteRequest', 'GetMessagesViewsRequest') 408 IMPORT_NAME_ATTR telethon.tl.functions.messages 410 IMPORT_FROM GetBotCallbackAnswerRequest 412 STORE_NAME GetBotCallbackAnswerRequest 414 IMPORT_FROM GetHistoryRequest 416 STORE_NAME GetHistoryRequest 418 IMPORT_FROM ImportChatInviteRequest 420 STORE_NAME ImportChatInviteRequest 422 IMPORT_FROM GetMessagesViewsRequest 424 STORE_NAME GetMessagesViewsRequest 426 POP_TOP L. 22 428 LOAD_CONST 0 430 LOAD_CONST ('datetime', 'timedelta') 432 IMPORT_NAME datetime 434 IMPORT_FROM datetime 436 STORE_NAME datetime 438 IMPORT_FROM timedelta 440 STORE_NAME timedelta 442 POP_TOP L. 23 444 LOAD_CONST 0 446 LOAD_CONST ('Fore', 'Style', 'init') 448 IMPORT_NAME colorama 450 IMPORT_FROM Fore 452 STORE_NAME Fore 454 IMPORT_FROM Style 456 STORE_NAME Style 458 IMPORT_FROM init 460 STORE_NAME color_ama 462 POP_TOP L. 24 464 LOAD_CONST 0 466 LOAD_CONST ('BeautifulSoup',) 468 IMPORT_NAME bs4 470 IMPORT_FROM BeautifulSoup 472 STORE_NAME BeautifulSoup 474 POP_TOP L. 25 476 LOAD_CONST 0 478 LOAD_CONST None 480 IMPORT_NAME random 482 STORE_NAME random L. 26 484 LOAD_CONST 0 486 LOAD_CONST None 488 IMPORT_NAME socks 490 STORE_NAME socks L. 27 492 LOAD_NAME logging 494 LOAD_ATTR basicConfig 496 LOAD_NAME logging 498 LOAD_ATTR ERROR 500 LOAD_CONST ('level',) 502 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args' 504 POP_TOP L. 28 506 LOAD_NAME color_ama 508 LOAD_CONST True 510 LOAD_CONST ('autoreset',) 512 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args' 514 POP_TOP L. 29 516 LOAD_NAME os 518 LOAD_METHOD system 520 LOAD_NAME os 522 LOAD_ATTR name 524 LOAD_STR 'nt' 526 COMPARE_OP == 528_530 POP_JUMP_IF_FALSE 536 'to 536' 532 LOAD_STR 'cls' 534 JUMP_FORWARD 538 'to 538' 536_0 COME_FROM 528 '528' 536 LOAD_STR 'clear' 538_0 COME_FROM 534 '534' 538 CALL_METHOD_1 1 '' 540 POP_TOP L. 30 542 LOAD_STR '\n' L. 31 544 LOAD_NAME Style 546 LOAD_ATTR NORMAL 548 LOAD_NAME Fore 550 LOAD_ATTR MAGENTA 552 BINARY_ADD L. 30 554 FORMAT_VALUE 0 '' 556 LOAD_STR ' ____ ___ ___ _ _ ' L. 31 558 LOAD_NAME Fore 560 LOAD_ATTR GREEN L. 30 562 FORMAT_VALUE 0 '' 564 LOAD_STR ' ___ ___ \n' L. 31 566 LOAD_NAME Style 568 LOAD_ATTR NORMAL 570 LOAD_NAME Fore 572 LOAD_ATTR MAGENTA 574 BINARY_ADD L. 30 576 FORMAT_VALUE 0 '' 578 LOAD_STR ' / ___| / _ \\ |_ _| | \\ | | ' L. 31 580 LOAD_NAME Fore 582 LOAD_ATTR GREEN L. 30 584 FORMAT_VALUE 0 '' 586 LOAD_STR ' / _ \\ ( _ ) \n' L. 31 588 LOAD_NAME Style 590 LOAD_ATTR NORMAL 592 LOAD_NAME Fore 594 LOAD_ATTR MAGENTA 596 BINARY_ADD L. 30 598 FORMAT_VALUE 0 '' 600 LOAD_STR ' | | | | | | | | | \\| | ' L. 31 602 LOAD_NAME Fore 604 LOAD_ATTR GREEN L. 30 606 FORMAT_VALUE 0 '' 608 LOAD_STR '| (_) | / _ \\ \n' L. 31 610 LOAD_NAME Style 612 LOAD_ATTR NORMAL 614 LOAD_NAME Fore 616 LOAD_ATTR MAGENTA 618 BINARY_ADD L. 30 620 FORMAT_VALUE 0 '' 622 LOAD_STR ' | |___ | |_| | | | | |\\ | ' L. 31 624 LOAD_NAME Fore 626 LOAD_ATTR GREEN L. 30 628 FORMAT_VALUE 0 '' 630 LOAD_STR ' \\__, | | (_) |\n' L. 31 632 LOAD_NAME Style 634 LOAD_ATTR NORMAL 636 LOAD_NAME Fore 638 LOAD_ATTR MAGENTA 640 BINARY_ADD L. 30 642 FORMAT_VALUE 0 '' 644 LOAD_STR ' \\____| \\___/ |___| |_| \\_| ' L. 31 646 LOAD_NAME Fore 648 LOAD_ATTR GREEN L. 30 650 FORMAT_VALUE 0 '' 652 LOAD_STR ' /_/ \\___/ \n' L. 36 654 LOAD_NAME Fore 656 LOAD_ATTR BLUE L. 30 658 FORMAT_VALUE 0 '' 660 LOAD_STR ' Edit By' L. 36 662 LOAD_NAME Style 664 LOAD_ATTR DIM 666 LOAD_NAME Fore 668 LOAD_ATTR RED 670 BINARY_ADD L. 30 672 FORMAT_VALUE 0 '' 674 LOAD_STR ':' L. 36 676 LOAD_NAME Fore 678 LOAD_ATTR YELLOW L. 30 680 FORMAT_VALUE 0 '' 682 LOAD_STR ' Abbas Bachari ' L. 36 684 LOAD_NAME Fore 686 LOAD_ATTR BLUE L. 30 688 FORMAT_VALUE 0 '' 690 LOAD_STR 'Version' L. 36 692 LOAD_NAME Fore 694 LOAD_ATTR YELLOW L. 30 696 FORMAT_VALUE 0 '' 698 LOAD_STR ' 7.9\n' L. 37 700 LOAD_NAME Style 702 LOAD_ATTR NORMAL 704 LOAD_NAME Fore 706 LOAD_ATTR RED 708 BINARY_ADD L. 30 710 FORMAT_VALUE 0 '' 712 LOAD_STR '==============================================\n' L. 38 714 LOAD_NAME Style 716 LOAD_ATTR BRIGHT 718 LOAD_NAME Fore 720 LOAD_ATTR GREEN 722 BINARY_ADD L. 30 724 FORMAT_VALUE 0 '' 726 LOAD_STR 'Sponsor Channel ' L. 36 728 LOAD_NAME Style 730 LOAD_ATTR DIM 732 LOAD_NAME Fore 734 LOAD_ATTR RED 736 BINARY_ADD L. 30 738 FORMAT_VALUE 0 '' 740 LOAD_STR ':' L. 38 742 LOAD_NAME Style 744 LOAD_ATTR RESET_ALL L. 30 746 FORMAT_VALUE 0 '' 748 LOAD_STR ' @COIN98' 750 BUILD_STRING_39 39 752 STORE_NAME banner L. 41 754 LOAD_CONST 799906641 756 LOAD_STR 'BitcoinClick_bot' 758 LOAD_STR '/start 2fv0' 760 LOAD_CONST ('id', 'username', 'start') 762 BUILD_CONST_KEY_MAP_3 3 L. 42 764 LOAD_CONST 741849360 766 LOAD_STR 'Litecoin_click_bot' 768 LOAD_STR '/start aZYG' 770 LOAD_CONST ('id', 'username', 'start') 772 BUILD_CONST_KEY_MAP_3 3 L. 43 774 LOAD_CONST 715510199 776 LOAD_STR 'Dogecoin_click_bot' 778 LOAD_STR '/start ljSP' 780 LOAD_CONST ('id', 'username', 'start') 782 BUILD_CONST_KEY_MAP_3 3 L. 44 784 LOAD_CONST 687127269 786 LOAD_STR 'BCH_clickbot' 788 LOAD_STR '/start 3su3' 790 LOAD_CONST ('id', 'username', 'start') 792 BUILD_CONST_KEY_MAP_3 3 L. 45 794 LOAD_CONST 850081470 796 LOAD_STR 'Zcash_click_bot' 798 LOAD_STR '/start 6qKG' 800 LOAD_CONST ('id', 'username', 'start') 802 BUILD_CONST_KEY_MAP_3 3 L. 40 804 LOAD_CONST ('BTC', 'LTC', 'DOGE', 'BCH', 'ZEC') 806 BUILD_CONST_KEY_MAP_5 5 808 STORE_NAME Symbols L. 48 810 LOAD_CONST 197597 812 STORE_GLOBAL api_id L. 49 814 LOAD_STR '227647fae1d2a3a1419bdd527337c87d' 816 STORE_GLOBAL api_hash L. 50 818 LOAD_STR 'RGVhciB1c2VyLCB5b3UgbmVlZCB0byBzdWJzY3JpYmUgdG8gdGhlCiAgICBmb2xsb3dpbmcgVGVsZWdyYW0gY2hhbm5lbBtbMTswbQogICAgQ2hhbm5lbCBJRCA6IEBDT0lOOTgNCgoK' 820 STORE_NAME MSG L. 51 822 LOAD_NAME requests 824 LOAD_METHOD session 826 CALL_METHOD_0 0 '' 828 STORE_GLOBAL session L. 52 830 LOAD_STR 'User-Agent' 832 LOAD_STR 'Mozilla/5.0 (Linux; Android 9; moto g(7) play) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.111 Mobile Safari/537.36' 834 BUILD_MAP_1 1 836 STORE_NAME ua L. 53 838 LOAD_STR '' 840 STORE_GLOBAL dogeclick_channel L. 54 842 LOAD_STR '' 844 STORE_GLOBAL phone_number L. 55 846 LOAD_STR '' 848 STORE_GLOBAL Wallet L. 56 850 LOAD_STR '' 852 STORE_GLOBAL isVisit L. 57 854 LOAD_STR '' 856 STORE_GLOBAL isJoin L. 58 858 LOAD_STR '' 860 STORE_GLOBAL isBotMs L. 59 862 LOAD_STR '' 864 STORE_GLOBAL isLoop L. 60 866 LOAD_STR '' 868 STORE_GLOBAL Balance L. 61 870 BUILD_MAP_0 0 872 STORE_GLOBAL SET L. 62 874 LOAD_STR '' 876 STORE_GLOBAL coin L. 63 878 LOAD_CONST 0 880 STORE_NAME INDEX L. 64 882 LOAD_CONST () 884 STORE_GLOBAL NUM L. 65 886 LOAD_STR 'SIT' 888 STORE_GLOBAL stat L. 66 890 LOAD_CONST None 892 STORE_GLOBAL captcha L. 67 894 LOAD_CONST None 896 STORE_GLOBAL sit L. 68 898 LOAD_STR '' 900 STORE_GLOBAL message L. 69 902 LOAD_STR '' 904 STORE_GLOBAL channel_name L. 70 906 LOAD_CONST None 908 STORE_GLOBAL channel_id L. 71 910 LOAD_STR '' 912 STORE_GLOBAL bot_name L. 72 914 LOAD_STR '' 916 STORE_GLOBAL COMENT L. 73 918 LOAD_CONST 7.85 920 STORE_NAME APVER L. 74 922 BUILD_LIST_0 0 924 STORE_GLOBAL BOTS L. 75 926 BUILD_LIST_0 0 928 STORE_GLOBAL CHNS L. 76 930 BUILD_LIST_0 0 932 STORE_GLOBAL GRPS L. 77 934 BUILD_LIST_0 0 936 STORE_GLOBAL USRS L. 78 938 LOAD_CONST 0 940 STORE_GLOBAL LIMIT L. 79 942 LOAD_STR '' 944 STORE_GLOBAL LICE L. 80 946 LOAD_STR '' 948 STORE_GLOBAL MASE L. 81 950 LOAD_CONST 0 952 STORE_GLOBAL VER L. 82 954 LOAD_CONST 0 956 STORE_GLOBAL mass_id L. 83 958 LOAD_CONST None 960 STORE_GLOBAL click_data L. 84 962 LOAD_STR '' 964 STORE_GLOBAL VIPMES L. 85 966 LOAD_CONST 0 968 STORE_GLOBAL JNUM L. 86 970 LOAD_CONST 0 972 STORE_GLOBAL ERN L. 87 974 LOAD_CONST 0 976 STORE_GLOBAL WIWNUM L. 88 978 LOAD_CONST 0 980 STORE_GLOBAL WIWCON L. 89 982 LOAD_CONST 0 984 STORE_NAME LOPS L. 90 986 BUILD_MAP_0 0 988 STORE_GLOBAL SETIN L. 91 990 BUILD_LIST_0 0 992 STORE_GLOBAL PLOGINS L. 92 994 BUILD_MAP_0 0 996 STORE_GLOBAL MESEGS L. 93 998 BUILD_MAP_0 0 1000 STORE_GLOBAL ADMESEG L. 94 1002 LOAD_STR 'YES' 1004 STORE_GLOBAL ADD L. 95 1006 LOAD_CONST None 1008 STORE_GLOBAL MINWD L. 96 1010 LOAD_STR 'NUMBERS.txt' 1012 STORE_GLOBAL path L. 97 1014 LOAD_CONST None 1016 STORE_GLOBAL Proxy L. 99 1018 LOAD_CODE <code_object isNumber> 1020 LOAD_STR 'isNumber' 1022 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1024 STORE_NAME isNumber L. 105 1026 LOAD_CODE <code_object request> 1028 LOAD_STR 'request' 1030 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1032 STORE_NAME request L. 131 1034 LOAD_CODE <code_object GetIP> 1036 LOAD_STR 'GetIP' 1038 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1040 STORE_NAME GetIP L. 145 1042 LOAD_CODE <code_object SEND_CONTACT> 1044 LOAD_STR 'SEND_CONTACT' 1046 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1048 STORE_NAME SEND_CONTACT L. 159 1050 LOAD_CODE <code_object send_messege_to_members> 1052 LOAD_STR 'send_messege_to_members' 1054 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1056 STORE_NAME send_messege_to_members L. 203 1058 LOAD_CODE <code_object CHNUM> 1060 LOAD_STR 'CHNUM' 1062 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1064 STORE_NAME CHNUM L. 211 1066 LOAD_CODE <code_object numbers> 1068 LOAD_STR 'numbers' 1070 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1072 STORE_NAME numbers L. 272 1074 LOAD_CODE <code_object AnonsRobot> 1076 LOAD_STR 'AnonsRobot' 1078 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1080 STORE_NAME AnonsRobot L. 285 1082 LOAD_CODE <code_object START_PLOGINS> 1084 LOAD_STR 'START_PLOGINS' 1086 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1088 STORE_NAME START_PLOGINS L. 382 1090 LOAD_CODE <code_object SEND_PLOGIN_MESSEGE> 1092 LOAD_STR 'SEND_PLOGIN_MESSEGE' 1094 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1096 STORE_NAME SEND_PLOGIN_MESSEGE L. 403 1098 LOAD_CODE <code_object SEND_GROUP_MESSEGE> 1100 LOAD_STR 'SEND_GROUP_MESSEGE' 1102 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1104 STORE_NAME SEND_GROUP_MESSEGE L. 435 1106 LOAD_CODE <code_object Settings> 1108 LOAD_STR 'Settings' 1110 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1112 STORE_NAME Settings L. 569 1114 LOAD_CODE <code_object SEND_TO_PRIVAIT_GROUP> 1116 LOAD_STR 'SEND_TO_PRIVAIT_GROUP' 1118 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1120 STORE_NAME SEND_TO_PRIVAIT_GROUP L. 611 1122 LOAD_CODE <code_object Creat_user> 1124 LOAD_STR 'Creat_user' 1126 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1128 STORE_NAME Creat_user L. 617 1130 LOAD_CODE <code_object Add_Plogin> 1132 LOAD_STR 'Add_Plogin' 1134 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1136 STORE_NAME Add_Plogin L. 629 1138 LOAD_CODE <code_object Delete_Plogin> 1140 LOAD_STR 'Delete_Plogin' 1142 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1144 STORE_NAME Delete_Plogin L. 635 1146 LOAD_CODE <code_object GetLoop> 1148 LOAD_STR 'GetLoop' 1150 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1152 STORE_NAME GetLoop L. 648 1154 LOAD_CODE <code_object Add_Channel> 1156 LOAD_STR 'Add_Channel' 1158 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1160 STORE_NAME Add_Channel L. 663 1162 LOAD_STR 'NO' 1164 STORE_GLOBAL ISWID L. 664 1166 LOAD_CODE <code_object Delete_Cannels> 1168 LOAD_STR 'Delete_Cannels' 1170 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1172 STORE_NAME Delete_Cannels L. 711 1174 LOAD_CODE <code_object reset> 1176 LOAD_STR 'reset' 1178 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1180 STORE_NAME reset L. 736 1182 LOAD_CODE <code_object maseg> 1184 LOAD_STR 'maseg' 1186 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1188 STORE_NAME maseg L. 740 1190 LOAD_CODE <code_object WAIT> 1192 LOAD_STR 'WAIT' 1194 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1196 STORE_NAME WAIT L. 777 1198 LOAD_CODE <code_object FloodTimer> 1200 LOAD_STR 'FloodTimer' 1202 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1204 STORE_NAME FloodTimer L. 787 1206 LOAD_CODE <code_object sleepTime> 1208 LOAD_STR 'sleepTime' 1210 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1212 STORE_NAME sleepTime L. 806 1214 LOAD_CODE <code_object JoinTimer> 1216 LOAD_STR 'JoinTimer' 1218 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1220 STORE_NAME JoinTimer L. 816 1222 LOAD_CODE <code_object GET_USERNAME> 1224 LOAD_STR 'GET_USERNAME' 1226 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1228 STORE_NAME GET_USERNAME L. 839 1230 LOAD_CODE <code_object forward_messages> 1232 LOAD_STR 'forward_messages' 1234 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1236 STORE_NAME forward_messages L. 854 1238 LOAD_CODE <code_object start> 1240 LOAD_STR 'start' 1242 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1244 STORE_NAME start L. 997 1246 LOAD_CODE <code_object main> 1248 LOAD_STR 'main' 1250 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1252 STORE_NAME main 1254_0 COME_FROM 1756 '1756' 1254_1 COME_FROM 1752 '1752' 1254_2 COME_FROM 1718 '1718' L.1293 1254_1256 SETUP_FINALLY 1720 'to 1720' L.1294 1258 LOAD_NAME asyncio 1260 LOAD_METHOD get_event_loop 1262 CALL_METHOD_0 0 '' 1264 LOAD_METHOD run_until_complete 1266 LOAD_NAME numbers 1268 CALL_FUNCTION_0 0 '' 1270 CALL_METHOD_1 1 '' 1272 POP_TOP L.1295 1274 LOAD_NAME LOPS 1276 LOAD_CONST 1 1278 INPLACE_ADD 1280 STORE_NAME LOPS L.1296 1282 LOAD_GLOBAL NUM 1284 GET_ITER 1286_0 COME_FROM 1328 '1328' 1286 FOR_ITER 1332 'to 1332' 1288 STORE_NAME i L.1297 1290 LOAD_NAME INDEX 1292 LOAD_CONST 1 1294 INPLACE_ADD 1296 STORE_NAME INDEX L.1298 1298 LOAD_CONST 0 1300 STORE_GLOBAL JNUM L.1299 1302 LOAD_NAME asyncio 1304 LOAD_METHOD get_event_loop 1306 CALL_METHOD_0 0 '' 1308 LOAD_METHOD run_until_complete 1310 LOAD_NAME main 1312 LOAD_NAME i 1314 CALL_FUNCTION_1 1 '' 1316 CALL_METHOD_1 1 '' 1318 POP_TOP L.1302 1320 LOAD_NAME sleep 1322 LOAD_CONST 3 1324 CALL_FUNCTION_1 1 '' 1326 POP_TOP 1328_1330 JUMP_BACK 1286 'to 1286' 1332_0 COME_FROM 1286 '1286' L.1303 1332 LOAD_NAME os 1334 LOAD_METHOD system 1336 LOAD_NAME os 1338 LOAD_ATTR name 1340 LOAD_STR 'nt' 1342 COMPARE_OP == 1344_1346 POP_JUMP_IF_FALSE 1352 'to 1352' 1348 LOAD_STR 'cls' 1350 JUMP_FORWARD 1354 'to 1354' 1352_0 COME_FROM 1344 '1344' 1352 LOAD_STR 'clear' 1354_0 COME_FROM 1350 '1350' 1354 CALL_METHOD_1 1 '' 1356 POP_TOP L.1304 1358 LOAD_NAME format 1360 LOAD_GLOBAL ERN 1362 LOAD_STR '.8f' 1364 CALL_FUNCTION_2 2 '' 1366 STORE_NAME ER L.1305 1368 LOAD_NAME format 1370 LOAD_GLOBAL WIWCON 1372 LOAD_STR '.8f' 1374 CALL_FUNCTION_2 2 '' 1376 STORE_NAME WC L.1306 1378 LOAD_GLOBAL LICE 1380 LOAD_STR 'VIP' 1382 COMPARE_OP == 1384_1386 POP_JUMP_IF_FALSE 1406 'to 1406' L.1307 1388 LOAD_GLOBAL LICE 1390 FORMAT_VALUE 0 '' 1392 LOAD_STR '(' 1394 LOAD_GLOBAL LIMIT 1396 FORMAT_VALUE 0 '' 1398 LOAD_STR ')' 1400 BUILD_STRING_4 4 1402 STORE_NAME VI 1404 JUMP_FORWARD 1428 'to 1428' 1406_0 COME_FROM 1384 '1384' L.1309 1406 LOAD_NAME Fore 1408 LOAD_ATTR RED 1410 FORMAT_VALUE 0 '' 1412 LOAD_GLOBAL LICE 1414 FORMAT_VALUE 0 '' 1416 LOAD_STR '(' 1418 LOAD_GLOBAL LIMIT 1420 FORMAT_VALUE 0 '' 1422 LOAD_STR ')' 1424 BUILD_STRING_5 5 1426 STORE_NAME VI 1428_0 COME_FROM 1404 '1404' L.1311 1428 LOAD_NAME print 1430 LOAD_NAME banner 1432 CALL_FUNCTION_1 1 '' 1434 POP_TOP L.1312 1436 LOAD_NAME print 1438 LOAD_NAME Fore 1440 LOAD_ATTR GREEN 1442 FORMAT_VALUE 0 '' 1444 LOAD_STR 'Script License ' 1446 LOAD_NAME Fore 1448 LOAD_ATTR RED 1450 FORMAT_VALUE 0 '' 1452 LOAD_STR ':' 1454 LOAD_NAME Fore 1456 LOAD_ATTR RESET 1458 FORMAT_VALUE 0 '' 1460 LOAD_STR ' ' 1462 LOAD_NAME VI 1464 FORMAT_VALUE 0 '' 1466 BUILD_STRING_7 7 1468 CALL_FUNCTION_1 1 '' 1470 POP_TOP L.1313 1472 LOAD_NAME print 1474 LOAD_NAME Fore 1476 LOAD_ATTR GREEN 1478 FORMAT_VALUE 0 '' 1480 LOAD_STR 'Total Loops ' 1482 LOAD_NAME Fore 1484 LOAD_ATTR RED 1486 FORMAT_VALUE 0 '' 1488 LOAD_STR ':' 1490 LOAD_NAME Fore 1492 LOAD_ATTR RESET 1494 FORMAT_VALUE 0 '' 1496 LOAD_STR ' ' 1498 LOAD_NAME LOPS 1500 FORMAT_VALUE 0 '' 1502 BUILD_STRING_7 7 1504 CALL_FUNCTION_1 1 '' 1506 POP_TOP L.1314 1508 LOAD_NAME print 1510 LOAD_NAME Fore 1512 LOAD_ATTR GREEN 1514 FORMAT_VALUE 0 '' 1516 LOAD_STR 'Total accounts ' 1518 LOAD_NAME Fore 1520 LOAD_ATTR RED 1522 FORMAT_VALUE 0 '' 1524 LOAD_STR ':' 1526 LOAD_NAME Fore 1528 LOAD_ATTR RESET 1530 FORMAT_VALUE 0 '' 1532 LOAD_STR ' ' 1534 LOAD_NAME len 1536 LOAD_GLOBAL NUM 1538 CALL_FUNCTION_1 1 '' 1540 FORMAT_VALUE 0 '' 1542 BUILD_STRING_7 7 1544 CALL_FUNCTION_1 1 '' 1546 POP_TOP L.1315 1548 LOAD_NAME print 1550 LOAD_NAME Fore 1552 LOAD_ATTR GREEN 1554 FORMAT_VALUE 0 '' 1556 LOAD_STR 'accounts withdraw ' 1558 LOAD_NAME Fore 1560 LOAD_ATTR RED 1562 FORMAT_VALUE 0 '' 1564 LOAD_STR ':' 1566 LOAD_NAME Fore 1568 LOAD_ATTR RESET 1570 FORMAT_VALUE 0 '' 1572 LOAD_STR ' ' 1574 LOAD_GLOBAL WIWNUM 1576 FORMAT_VALUE 0 '' 1578 LOAD_STR ' account' 1580 BUILD_STRING_8 8 1582 CALL_FUNCTION_1 1 '' 1584 POP_TOP L.1316 1586 LOAD_NAME print 1588 LOAD_NAME Fore 1590 LOAD_ATTR GREEN 1592 FORMAT_VALUE 0 '' 1594 LOAD_STR 'Total withdraw ' 1596 LOAD_NAME Fore 1598 LOAD_ATTR RED 1600 FORMAT_VALUE 0 '' 1602 LOAD_STR ':' 1604 LOAD_NAME Fore 1606 LOAD_ATTR RESET 1608 FORMAT_VALUE 0 '' 1610 LOAD_STR ' ' 1612 LOAD_NAME WC 1614 FORMAT_VALUE 0 '' 1616 LOAD_STR ' ' 1618 LOAD_GLOBAL coin 1620 FORMAT_VALUE 0 '' 1622 BUILD_STRING_9 9 1624 CALL_FUNCTION_1 1 '' 1626 POP_TOP L.1317 1628 LOAD_NAME print 1630 LOAD_NAME Fore 1632 LOAD_ATTR GREEN 1634 FORMAT_VALUE 0 '' 1636 LOAD_STR 'Total balance ' 1638 LOAD_NAME Fore 1640 LOAD_ATTR RED 1642 FORMAT_VALUE 0 '' 1644 LOAD_STR ':' 1646 LOAD_NAME Fore 1648 LOAD_ATTR RESET 1650 FORMAT_VALUE 0 '' 1652 LOAD_STR ' ' 1654 LOAD_NAME ER 1656 FORMAT_VALUE 0 '' 1658 LOAD_STR ' ' 1660 LOAD_GLOBAL coin 1662 FORMAT_VALUE 0 '' 1664 BUILD_STRING_9 9 1666 CALL_FUNCTION_1 1 '' 1668 POP_TOP L.1318 1670 LOAD_NAME print 1672 LOAD_STR '\n' 1674 CALL_FUNCTION_1 1 '' 1676 POP_TOP L.1319 1678 LOAD_GLOBAL SET 1680 LOAD_STR 'AUTOLOOP' 1682 BINARY_SUBSCR 1684 LOAD_METHOD upper 1686 CALL_METHOD_0 0 '' 1688 LOAD_STR 'YES' 1690 COMPARE_OP != 1692_1694 POP_JUMP_IF_FALSE 1702 'to 1702' L.1320 1696 POP_BLOCK 1698_1700 BREAK_LOOP 1760 'to 1760' 1702_0 COME_FROM 1692 '1692' L.1322 1702 LOAD_NAME sleepTime 1704 CALL_FUNCTION_0 0 '' 1706 POP_TOP L.1323 1708 LOAD_CONST 0 1710 STORE_GLOBAL ERN L.1324 1712 LOAD_CONST 0 1714 STORE_NAME INDEX 1716 POP_BLOCK 1718 JUMP_BACK 1254 'to 1254' 1720_0 COME_FROM_FINALLY 1254 '1254' L.1326 1720 DUP_TOP 1722 LOAD_NAME KeyboardInterrupt 1724_1726 <121> 1754 '' 1728 POP_TOP 1730 POP_TOP 1732 POP_TOP L.1327 1734 LOAD_NAME print 1736 LOAD_STR '' 1738 CALL_FUNCTION_1 1 '' 1740 POP_TOP L.1328 1742 LOAD_NAME sys 1744 LOAD_METHOD exit 1746 CALL_METHOD_0 0 '' 1748 POP_TOP 1750 POP_EXCEPT 1752 JUMP_BACK 1254 'to 1254' 1754 <48> 1756_1758 JUMP_BACK 1254 'to 1254' 1760_0 COME_FROM 1698 '1698' Parse error at or near `DUP_TOP' instruction at offset 84 ``` i also attached : .`pyc `file , decompiled results and `pydisasm `results [decompiled_output.zip](https://github.com/rocky/python-decompile3/files/5471265/decompiled_output.zip) [pycfile.zip](https://github.com/rocky/python-decompile3/files/5471266/pycfile.zip) [pydisasm.zip](https://github.com/rocky/python-decompile3/files/5471267/pydisasm.zip)
1.0
Parse error at or near `DUP_TOP' (python 3.8) - When I decompiled the pyc, got the following error messages. OS : win 10 Python 3.8.6 uncompyle6 : 3.7.4 decompyle3 : 3.3.2 pydisasm version 5.0.4 Python bytecode 3.8 (3413) Source code size mod 2**32 ``` Instruction context: L. 5 52 LOAD_NAME list 54 LOAD_GENEXPR '<code_object <genexpr>>' 56 LOAD_STR '<genexpr>' 58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 60 LOAD_GENEXPR '<code_object <genexpr>>' 62 LOAD_STR '<genexpr>' 64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 66 LOAD_NAME f_in 68 GET_ITER 70 CALL_FUNCTION_1 1 '' 72 GET_ITER 74 CALL_FUNCTION_1 1 '' 76 CALL_FUNCTION_1 1 '' 78 STORE_NAME requirements 80 POP_BLOCK 82 LOAD_CONST None -> 84 DUP_TOP 86 DUP_TOP 88 CALL_FUNCTION_3 3 '' 90 POP_TOP 92 JUMP_FORWARD 110 'to 110' 94_0 COME_FROM_WITH 48 '48' 94 <49> 96 POP_JUMP_IF_TRUE 100 'to 100' 98 <48> 100_0 COME_FROM 96 '96' 100 POP_TOP 102 POP_TOP 104 POP_TOP 106 POP_EXCEPT 108 POP_TOP 110_0 COME_FROM 92 '92' # file start.pyc # --- This code section failed: --- L. 2 0 LOAD_CONST 0 2 LOAD_CONST None 4 IMPORT_NAME re 6 STORE_NAME re 8 LOAD_CONST 0 10 LOAD_CONST None 12 IMPORT_NAME os 14 STORE_NAME os 16 LOAD_CONST 0 18 LOAD_CONST None 20 IMPORT_NAME sys 22 STORE_NAME sys 24 LOAD_CONST 0 26 LOAD_CONST None 28 IMPORT_NAME pkg_resources 30 STORE_NAME pkg_resources 32 LOAD_CONST 0 34 LOAD_CONST None 36 IMPORT_NAME random 38 STORE_NAME random L. 3 40 SETUP_FINALLY 228 'to 228' L. 4 42 LOAD_NAME open 44 LOAD_STR 'requirements.txt' 46 CALL_FUNCTION_1 1 '' 48 SETUP_WITH 94 'to 94' 50 STORE_NAME f_in L. 5 52 LOAD_NAME list 54 LOAD_GENEXPR '<code_object <genexpr>>' 56 LOAD_STR '<genexpr>' 58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 60 LOAD_GENEXPR '<code_object <genexpr>>' 62 LOAD_STR '<genexpr>' 64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 66 LOAD_NAME f_in 68 GET_ITER 70 CALL_FUNCTION_1 1 '' 72 GET_ITER 74 CALL_FUNCTION_1 1 '' 76 CALL_FUNCTION_1 1 '' 78 STORE_NAME requirements 80 POP_BLOCK 82 LOAD_CONST None 84 DUP_TOP 86 DUP_TOP 88 CALL_FUNCTION_3 3 '' 90 POP_TOP 92 JUMP_FORWARD 110 'to 110' 94_0 COME_FROM_WITH 48 '48' 94 <49> 96 POP_JUMP_IF_TRUE 100 'to 100' 98 <48> 100_0 COME_FROM 96 '96' 100 POP_TOP 102 POP_TOP 104 POP_TOP 106 POP_EXCEPT 108 POP_TOP 110_0 COME_FROM 92 '92' L. 6 110 LOAD_NAME requirements 112 GET_ITER 114_0 COME_FROM 222 '222' 114_1 COME_FROM 210 '210' 114_2 COME_FROM 132 '132' 114 FOR_ITER 224 'to 224' 116 STORE_NAME require L. 7 118 SETUP_FINALLY 134 'to 134' L. 8 120 LOAD_NAME pkg_resources 122 LOAD_METHOD require 124 LOAD_NAME require 126 CALL_METHOD_1 1 '' 128 POP_TOP 130 POP_BLOCK 132 JUMP_BACK 114 'to 114' 134_0 COME_FROM_FINALLY 118 '118' L. 9 134 DUP_TOP 136 LOAD_NAME Exception 138 <121> 220 '' 140 POP_TOP 142 STORE_NAME e 144 POP_TOP 146 SETUP_FINALLY 212 'to 212' L. 10 148 LOAD_NAME re 150 LOAD_METHOD search 152 LOAD_STR "\\'(.*?)\\'" 154 LOAD_NAME str 156 LOAD_NAME e 158 CALL_FUNCTION_1 1 '' 160 CALL_METHOD_2 2 '' 162 LOAD_METHOD group 164 LOAD_CONST 1 166 CALL_METHOD_1 1 '' 168 STORE_NAME a L. 11 170 LOAD_NAME print 172 LOAD_STR 'Installing ' 174 LOAD_NAME a 176 BINARY_ADD 178 LOAD_STR ' ...' 180 BINARY_ADD 182 CALL_FUNCTION_1 1 '' 184 POP_TOP L. 12 186 LOAD_NAME os 188 LOAD_METHOD system 190 LOAD_STR 'pip install ' 192 LOAD_NAME a 194 BINARY_ADD 196 CALL_METHOD_1 1 '' 198 POP_TOP 200 POP_BLOCK 202 POP_EXCEPT 204 LOAD_CONST None 206 STORE_NAME e 208 DELETE_NAME e 210 JUMP_BACK 114 'to 114' 212_0 COME_FROM_FINALLY 146 '146' 212 LOAD_CONST None 214 STORE_NAME e 216 DELETE_NAME e 218 <48> 220 <48> 222 JUMP_BACK 114 'to 114' 224_0 COME_FROM 114 '114' 224 POP_BLOCK 226 JUMP_FORWARD 280 'to 280' 228_0 COME_FROM_FINALLY 40 '40' L. 13 228 DUP_TOP 230 LOAD_NAME Exception 232_234 <121> 278 '' 236 POP_TOP 238 STORE_NAME e 240 POP_TOP 242 SETUP_FINALLY 270 'to 270' L. 14 244 LOAD_NAME print 246 LOAD_STR 'installation error: ' 248 LOAD_NAME e 250 FORMAT_VALUE 0 '' 252 BUILD_STRING_2 2 254 CALL_FUNCTION_1 1 '' 256 POP_TOP 258 POP_BLOCK 260 POP_EXCEPT 262 LOAD_CONST None 264 STORE_NAME e 266 DELETE_NAME e 268 JUMP_FORWARD 280 'to 280' 270_0 COME_FROM_FINALLY 242 '242' 270 LOAD_CONST None 272 STORE_NAME e 274 DELETE_NAME e 276 <48> 278 <48> 280_0 COME_FROM 268 '268' 280_1 COME_FROM 226 '226' L. 16 280 LOAD_CONST 0 282 LOAD_CONST None 284 IMPORT_NAME asyncio 286 STORE_NAME asyncio 288 LOAD_CONST 0 290 LOAD_CONST None 292 IMPORT_NAME logging 294 STORE_NAME logging 296 LOAD_CONST 0 298 LOAD_CONST None 300 IMPORT_NAME requests 302 STORE_NAME requests 304 LOAD_CONST 0 306 LOAD_CONST None 308 IMPORT_NAME sqlite3 310 STORE_NAME sqlite3 312 LOAD_CONST 0 314 LOAD_CONST None 316 IMPORT_NAME base64 318 STORE_NAME base64 L. 17 320 LOAD_CONST 0 322 LOAD_CONST ('sleep', 'strptime') 324 IMPORT_NAME time 326 IMPORT_FROM sleep 328 STORE_NAME sleep 330 IMPORT_FROM strptime 332 STORE_NAME strptime 334 POP_TOP L. 18 336 LOAD_CONST 0 338 LOAD_CONST ('KeyboardButtonUrl', 'MessageMediaContact', 'ReplyInlineMarkup', 'UpdateShortMessage') 340 IMPORT_NAME_ATTR telethon.tl.types 342 IMPORT_FROM KeyboardButtonUrl 344 STORE_NAME KeyboardButtonUrl 346 IMPORT_FROM MessageMediaContact 348 STORE_NAME MessageMediaContact 350 IMPORT_FROM ReplyInlineMarkup 352 STORE_NAME ReplyInlineMarkup 354 IMPORT_FROM UpdateShortMessage 356 STORE_NAME UpdateShortMessage 358 POP_TOP L. 19 360 LOAD_CONST 0 362 LOAD_CONST ('TelegramClient', 'client', 'errors', 'events', 'functions', 'connection') 364 IMPORT_NAME telethon 366 IMPORT_FROM TelegramClient 368 STORE_NAME TelegramClient 370 IMPORT_FROM client 372 STORE_GLOBAL client 374 IMPORT_FROM errors 376 STORE_NAME errors 378 IMPORT_FROM events 380 STORE_NAME events 382 IMPORT_FROM functions 384 STORE_NAME functions 386 IMPORT_FROM connection 388 STORE_NAME connection 390 POP_TOP L. 20 392 LOAD_CONST 0 394 LOAD_CONST ('JoinChannelRequest',) 396 IMPORT_NAME_ATTR telethon.tl.functions.channels 398 IMPORT_FROM JoinChannelRequest 400 STORE_NAME JoinChannelRequest 402 POP_TOP L. 21 404 LOAD_CONST 0 406 LOAD_CONST ('GetBotCallbackAnswerRequest', 'GetHistoryRequest', 'ImportChatInviteRequest', 'GetMessagesViewsRequest') 408 IMPORT_NAME_ATTR telethon.tl.functions.messages 410 IMPORT_FROM GetBotCallbackAnswerRequest 412 STORE_NAME GetBotCallbackAnswerRequest 414 IMPORT_FROM GetHistoryRequest 416 STORE_NAME GetHistoryRequest 418 IMPORT_FROM ImportChatInviteRequest 420 STORE_NAME ImportChatInviteRequest 422 IMPORT_FROM GetMessagesViewsRequest 424 STORE_NAME GetMessagesViewsRequest 426 POP_TOP L. 22 428 LOAD_CONST 0 430 LOAD_CONST ('datetime', 'timedelta') 432 IMPORT_NAME datetime 434 IMPORT_FROM datetime 436 STORE_NAME datetime 438 IMPORT_FROM timedelta 440 STORE_NAME timedelta 442 POP_TOP L. 23 444 LOAD_CONST 0 446 LOAD_CONST ('Fore', 'Style', 'init') 448 IMPORT_NAME colorama 450 IMPORT_FROM Fore 452 STORE_NAME Fore 454 IMPORT_FROM Style 456 STORE_NAME Style 458 IMPORT_FROM init 460 STORE_NAME color_ama 462 POP_TOP L. 24 464 LOAD_CONST 0 466 LOAD_CONST ('BeautifulSoup',) 468 IMPORT_NAME bs4 470 IMPORT_FROM BeautifulSoup 472 STORE_NAME BeautifulSoup 474 POP_TOP L. 25 476 LOAD_CONST 0 478 LOAD_CONST None 480 IMPORT_NAME random 482 STORE_NAME random L. 26 484 LOAD_CONST 0 486 LOAD_CONST None 488 IMPORT_NAME socks 490 STORE_NAME socks L. 27 492 LOAD_NAME logging 494 LOAD_ATTR basicConfig 496 LOAD_NAME logging 498 LOAD_ATTR ERROR 500 LOAD_CONST ('level',) 502 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args' 504 POP_TOP L. 28 506 LOAD_NAME color_ama 508 LOAD_CONST True 510 LOAD_CONST ('autoreset',) 512 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args' 514 POP_TOP L. 29 516 LOAD_NAME os 518 LOAD_METHOD system 520 LOAD_NAME os 522 LOAD_ATTR name 524 LOAD_STR 'nt' 526 COMPARE_OP == 528_530 POP_JUMP_IF_FALSE 536 'to 536' 532 LOAD_STR 'cls' 534 JUMP_FORWARD 538 'to 538' 536_0 COME_FROM 528 '528' 536 LOAD_STR 'clear' 538_0 COME_FROM 534 '534' 538 CALL_METHOD_1 1 '' 540 POP_TOP L. 30 542 LOAD_STR '\n' L. 31 544 LOAD_NAME Style 546 LOAD_ATTR NORMAL 548 LOAD_NAME Fore 550 LOAD_ATTR MAGENTA 552 BINARY_ADD L. 30 554 FORMAT_VALUE 0 '' 556 LOAD_STR ' ____ ___ ___ _ _ ' L. 31 558 LOAD_NAME Fore 560 LOAD_ATTR GREEN L. 30 562 FORMAT_VALUE 0 '' 564 LOAD_STR ' ___ ___ \n' L. 31 566 LOAD_NAME Style 568 LOAD_ATTR NORMAL 570 LOAD_NAME Fore 572 LOAD_ATTR MAGENTA 574 BINARY_ADD L. 30 576 FORMAT_VALUE 0 '' 578 LOAD_STR ' / ___| / _ \\ |_ _| | \\ | | ' L. 31 580 LOAD_NAME Fore 582 LOAD_ATTR GREEN L. 30 584 FORMAT_VALUE 0 '' 586 LOAD_STR ' / _ \\ ( _ ) \n' L. 31 588 LOAD_NAME Style 590 LOAD_ATTR NORMAL 592 LOAD_NAME Fore 594 LOAD_ATTR MAGENTA 596 BINARY_ADD L. 30 598 FORMAT_VALUE 0 '' 600 LOAD_STR ' | | | | | | | | | \\| | ' L. 31 602 LOAD_NAME Fore 604 LOAD_ATTR GREEN L. 30 606 FORMAT_VALUE 0 '' 608 LOAD_STR '| (_) | / _ \\ \n' L. 31 610 LOAD_NAME Style 612 LOAD_ATTR NORMAL 614 LOAD_NAME Fore 616 LOAD_ATTR MAGENTA 618 BINARY_ADD L. 30 620 FORMAT_VALUE 0 '' 622 LOAD_STR ' | |___ | |_| | | | | |\\ | ' L. 31 624 LOAD_NAME Fore 626 LOAD_ATTR GREEN L. 30 628 FORMAT_VALUE 0 '' 630 LOAD_STR ' \\__, | | (_) |\n' L. 31 632 LOAD_NAME Style 634 LOAD_ATTR NORMAL 636 LOAD_NAME Fore 638 LOAD_ATTR MAGENTA 640 BINARY_ADD L. 30 642 FORMAT_VALUE 0 '' 644 LOAD_STR ' \\____| \\___/ |___| |_| \\_| ' L. 31 646 LOAD_NAME Fore 648 LOAD_ATTR GREEN L. 30 650 FORMAT_VALUE 0 '' 652 LOAD_STR ' /_/ \\___/ \n' L. 36 654 LOAD_NAME Fore 656 LOAD_ATTR BLUE L. 30 658 FORMAT_VALUE 0 '' 660 LOAD_STR ' Edit By' L. 36 662 LOAD_NAME Style 664 LOAD_ATTR DIM 666 LOAD_NAME Fore 668 LOAD_ATTR RED 670 BINARY_ADD L. 30 672 FORMAT_VALUE 0 '' 674 LOAD_STR ':' L. 36 676 LOAD_NAME Fore 678 LOAD_ATTR YELLOW L. 30 680 FORMAT_VALUE 0 '' 682 LOAD_STR ' Abbas Bachari ' L. 36 684 LOAD_NAME Fore 686 LOAD_ATTR BLUE L. 30 688 FORMAT_VALUE 0 '' 690 LOAD_STR 'Version' L. 36 692 LOAD_NAME Fore 694 LOAD_ATTR YELLOW L. 30 696 FORMAT_VALUE 0 '' 698 LOAD_STR ' 7.9\n' L. 37 700 LOAD_NAME Style 702 LOAD_ATTR NORMAL 704 LOAD_NAME Fore 706 LOAD_ATTR RED 708 BINARY_ADD L. 30 710 FORMAT_VALUE 0 '' 712 LOAD_STR '==============================================\n' L. 38 714 LOAD_NAME Style 716 LOAD_ATTR BRIGHT 718 LOAD_NAME Fore 720 LOAD_ATTR GREEN 722 BINARY_ADD L. 30 724 FORMAT_VALUE 0 '' 726 LOAD_STR 'Sponsor Channel ' L. 36 728 LOAD_NAME Style 730 LOAD_ATTR DIM 732 LOAD_NAME Fore 734 LOAD_ATTR RED 736 BINARY_ADD L. 30 738 FORMAT_VALUE 0 '' 740 LOAD_STR ':' L. 38 742 LOAD_NAME Style 744 LOAD_ATTR RESET_ALL L. 30 746 FORMAT_VALUE 0 '' 748 LOAD_STR ' @COIN98' 750 BUILD_STRING_39 39 752 STORE_NAME banner L. 41 754 LOAD_CONST 799906641 756 LOAD_STR 'BitcoinClick_bot' 758 LOAD_STR '/start 2fv0' 760 LOAD_CONST ('id', 'username', 'start') 762 BUILD_CONST_KEY_MAP_3 3 L. 42 764 LOAD_CONST 741849360 766 LOAD_STR 'Litecoin_click_bot' 768 LOAD_STR '/start aZYG' 770 LOAD_CONST ('id', 'username', 'start') 772 BUILD_CONST_KEY_MAP_3 3 L. 43 774 LOAD_CONST 715510199 776 LOAD_STR 'Dogecoin_click_bot' 778 LOAD_STR '/start ljSP' 780 LOAD_CONST ('id', 'username', 'start') 782 BUILD_CONST_KEY_MAP_3 3 L. 44 784 LOAD_CONST 687127269 786 LOAD_STR 'BCH_clickbot' 788 LOAD_STR '/start 3su3' 790 LOAD_CONST ('id', 'username', 'start') 792 BUILD_CONST_KEY_MAP_3 3 L. 45 794 LOAD_CONST 850081470 796 LOAD_STR 'Zcash_click_bot' 798 LOAD_STR '/start 6qKG' 800 LOAD_CONST ('id', 'username', 'start') 802 BUILD_CONST_KEY_MAP_3 3 L. 40 804 LOAD_CONST ('BTC', 'LTC', 'DOGE', 'BCH', 'ZEC') 806 BUILD_CONST_KEY_MAP_5 5 808 STORE_NAME Symbols L. 48 810 LOAD_CONST 197597 812 STORE_GLOBAL api_id L. 49 814 LOAD_STR '227647fae1d2a3a1419bdd527337c87d' 816 STORE_GLOBAL api_hash L. 50 818 LOAD_STR 'RGVhciB1c2VyLCB5b3UgbmVlZCB0byBzdWJzY3JpYmUgdG8gdGhlCiAgICBmb2xsb3dpbmcgVGVsZWdyYW0gY2hhbm5lbBtbMTswbQogICAgQ2hhbm5lbCBJRCA6IEBDT0lOOTgNCgoK' 820 STORE_NAME MSG L. 51 822 LOAD_NAME requests 824 LOAD_METHOD session 826 CALL_METHOD_0 0 '' 828 STORE_GLOBAL session L. 52 830 LOAD_STR 'User-Agent' 832 LOAD_STR 'Mozilla/5.0 (Linux; Android 9; moto g(7) play) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.111 Mobile Safari/537.36' 834 BUILD_MAP_1 1 836 STORE_NAME ua L. 53 838 LOAD_STR '' 840 STORE_GLOBAL dogeclick_channel L. 54 842 LOAD_STR '' 844 STORE_GLOBAL phone_number L. 55 846 LOAD_STR '' 848 STORE_GLOBAL Wallet L. 56 850 LOAD_STR '' 852 STORE_GLOBAL isVisit L. 57 854 LOAD_STR '' 856 STORE_GLOBAL isJoin L. 58 858 LOAD_STR '' 860 STORE_GLOBAL isBotMs L. 59 862 LOAD_STR '' 864 STORE_GLOBAL isLoop L. 60 866 LOAD_STR '' 868 STORE_GLOBAL Balance L. 61 870 BUILD_MAP_0 0 872 STORE_GLOBAL SET L. 62 874 LOAD_STR '' 876 STORE_GLOBAL coin L. 63 878 LOAD_CONST 0 880 STORE_NAME INDEX L. 64 882 LOAD_CONST () 884 STORE_GLOBAL NUM L. 65 886 LOAD_STR 'SIT' 888 STORE_GLOBAL stat L. 66 890 LOAD_CONST None 892 STORE_GLOBAL captcha L. 67 894 LOAD_CONST None 896 STORE_GLOBAL sit L. 68 898 LOAD_STR '' 900 STORE_GLOBAL message L. 69 902 LOAD_STR '' 904 STORE_GLOBAL channel_name L. 70 906 LOAD_CONST None 908 STORE_GLOBAL channel_id L. 71 910 LOAD_STR '' 912 STORE_GLOBAL bot_name L. 72 914 LOAD_STR '' 916 STORE_GLOBAL COMENT L. 73 918 LOAD_CONST 7.85 920 STORE_NAME APVER L. 74 922 BUILD_LIST_0 0 924 STORE_GLOBAL BOTS L. 75 926 BUILD_LIST_0 0 928 STORE_GLOBAL CHNS L. 76 930 BUILD_LIST_0 0 932 STORE_GLOBAL GRPS L. 77 934 BUILD_LIST_0 0 936 STORE_GLOBAL USRS L. 78 938 LOAD_CONST 0 940 STORE_GLOBAL LIMIT L. 79 942 LOAD_STR '' 944 STORE_GLOBAL LICE L. 80 946 LOAD_STR '' 948 STORE_GLOBAL MASE L. 81 950 LOAD_CONST 0 952 STORE_GLOBAL VER L. 82 954 LOAD_CONST 0 956 STORE_GLOBAL mass_id L. 83 958 LOAD_CONST None 960 STORE_GLOBAL click_data L. 84 962 LOAD_STR '' 964 STORE_GLOBAL VIPMES L. 85 966 LOAD_CONST 0 968 STORE_GLOBAL JNUM L. 86 970 LOAD_CONST 0 972 STORE_GLOBAL ERN L. 87 974 LOAD_CONST 0 976 STORE_GLOBAL WIWNUM L. 88 978 LOAD_CONST 0 980 STORE_GLOBAL WIWCON L. 89 982 LOAD_CONST 0 984 STORE_NAME LOPS L. 90 986 BUILD_MAP_0 0 988 STORE_GLOBAL SETIN L. 91 990 BUILD_LIST_0 0 992 STORE_GLOBAL PLOGINS L. 92 994 BUILD_MAP_0 0 996 STORE_GLOBAL MESEGS L. 93 998 BUILD_MAP_0 0 1000 STORE_GLOBAL ADMESEG L. 94 1002 LOAD_STR 'YES' 1004 STORE_GLOBAL ADD L. 95 1006 LOAD_CONST None 1008 STORE_GLOBAL MINWD L. 96 1010 LOAD_STR 'NUMBERS.txt' 1012 STORE_GLOBAL path L. 97 1014 LOAD_CONST None 1016 STORE_GLOBAL Proxy L. 99 1018 LOAD_CODE <code_object isNumber> 1020 LOAD_STR 'isNumber' 1022 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1024 STORE_NAME isNumber L. 105 1026 LOAD_CODE <code_object request> 1028 LOAD_STR 'request' 1030 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1032 STORE_NAME request L. 131 1034 LOAD_CODE <code_object GetIP> 1036 LOAD_STR 'GetIP' 1038 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1040 STORE_NAME GetIP L. 145 1042 LOAD_CODE <code_object SEND_CONTACT> 1044 LOAD_STR 'SEND_CONTACT' 1046 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1048 STORE_NAME SEND_CONTACT L. 159 1050 LOAD_CODE <code_object send_messege_to_members> 1052 LOAD_STR 'send_messege_to_members' 1054 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1056 STORE_NAME send_messege_to_members L. 203 1058 LOAD_CODE <code_object CHNUM> 1060 LOAD_STR 'CHNUM' 1062 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1064 STORE_NAME CHNUM L. 211 1066 LOAD_CODE <code_object numbers> 1068 LOAD_STR 'numbers' 1070 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1072 STORE_NAME numbers L. 272 1074 LOAD_CODE <code_object AnonsRobot> 1076 LOAD_STR 'AnonsRobot' 1078 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1080 STORE_NAME AnonsRobot L. 285 1082 LOAD_CODE <code_object START_PLOGINS> 1084 LOAD_STR 'START_PLOGINS' 1086 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1088 STORE_NAME START_PLOGINS L. 382 1090 LOAD_CODE <code_object SEND_PLOGIN_MESSEGE> 1092 LOAD_STR 'SEND_PLOGIN_MESSEGE' 1094 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1096 STORE_NAME SEND_PLOGIN_MESSEGE L. 403 1098 LOAD_CODE <code_object SEND_GROUP_MESSEGE> 1100 LOAD_STR 'SEND_GROUP_MESSEGE' 1102 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1104 STORE_NAME SEND_GROUP_MESSEGE L. 435 1106 LOAD_CODE <code_object Settings> 1108 LOAD_STR 'Settings' 1110 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1112 STORE_NAME Settings L. 569 1114 LOAD_CODE <code_object SEND_TO_PRIVAIT_GROUP> 1116 LOAD_STR 'SEND_TO_PRIVAIT_GROUP' 1118 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1120 STORE_NAME SEND_TO_PRIVAIT_GROUP L. 611 1122 LOAD_CODE <code_object Creat_user> 1124 LOAD_STR 'Creat_user' 1126 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1128 STORE_NAME Creat_user L. 617 1130 LOAD_CODE <code_object Add_Plogin> 1132 LOAD_STR 'Add_Plogin' 1134 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1136 STORE_NAME Add_Plogin L. 629 1138 LOAD_CODE <code_object Delete_Plogin> 1140 LOAD_STR 'Delete_Plogin' 1142 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1144 STORE_NAME Delete_Plogin L. 635 1146 LOAD_CODE <code_object GetLoop> 1148 LOAD_STR 'GetLoop' 1150 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1152 STORE_NAME GetLoop L. 648 1154 LOAD_CODE <code_object Add_Channel> 1156 LOAD_STR 'Add_Channel' 1158 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1160 STORE_NAME Add_Channel L. 663 1162 LOAD_STR 'NO' 1164 STORE_GLOBAL ISWID L. 664 1166 LOAD_CODE <code_object Delete_Cannels> 1168 LOAD_STR 'Delete_Cannels' 1170 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1172 STORE_NAME Delete_Cannels L. 711 1174 LOAD_CODE <code_object reset> 1176 LOAD_STR 'reset' 1178 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1180 STORE_NAME reset L. 736 1182 LOAD_CODE <code_object maseg> 1184 LOAD_STR 'maseg' 1186 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1188 STORE_NAME maseg L. 740 1190 LOAD_CODE <code_object WAIT> 1192 LOAD_STR 'WAIT' 1194 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1196 STORE_NAME WAIT L. 777 1198 LOAD_CODE <code_object FloodTimer> 1200 LOAD_STR 'FloodTimer' 1202 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1204 STORE_NAME FloodTimer L. 787 1206 LOAD_CODE <code_object sleepTime> 1208 LOAD_STR 'sleepTime' 1210 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1212 STORE_NAME sleepTime L. 806 1214 LOAD_CODE <code_object JoinTimer> 1216 LOAD_STR 'JoinTimer' 1218 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1220 STORE_NAME JoinTimer L. 816 1222 LOAD_CODE <code_object GET_USERNAME> 1224 LOAD_STR 'GET_USERNAME' 1226 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1228 STORE_NAME GET_USERNAME L. 839 1230 LOAD_CODE <code_object forward_messages> 1232 LOAD_STR 'forward_messages' 1234 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1236 STORE_NAME forward_messages L. 854 1238 LOAD_CODE <code_object start> 1240 LOAD_STR 'start' 1242 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1244 STORE_NAME start L. 997 1246 LOAD_CODE <code_object main> 1248 LOAD_STR 'main' 1250 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures' 1252 STORE_NAME main 1254_0 COME_FROM 1756 '1756' 1254_1 COME_FROM 1752 '1752' 1254_2 COME_FROM 1718 '1718' L.1293 1254_1256 SETUP_FINALLY 1720 'to 1720' L.1294 1258 LOAD_NAME asyncio 1260 LOAD_METHOD get_event_loop 1262 CALL_METHOD_0 0 '' 1264 LOAD_METHOD run_until_complete 1266 LOAD_NAME numbers 1268 CALL_FUNCTION_0 0 '' 1270 CALL_METHOD_1 1 '' 1272 POP_TOP L.1295 1274 LOAD_NAME LOPS 1276 LOAD_CONST 1 1278 INPLACE_ADD 1280 STORE_NAME LOPS L.1296 1282 LOAD_GLOBAL NUM 1284 GET_ITER 1286_0 COME_FROM 1328 '1328' 1286 FOR_ITER 1332 'to 1332' 1288 STORE_NAME i L.1297 1290 LOAD_NAME INDEX 1292 LOAD_CONST 1 1294 INPLACE_ADD 1296 STORE_NAME INDEX L.1298 1298 LOAD_CONST 0 1300 STORE_GLOBAL JNUM L.1299 1302 LOAD_NAME asyncio 1304 LOAD_METHOD get_event_loop 1306 CALL_METHOD_0 0 '' 1308 LOAD_METHOD run_until_complete 1310 LOAD_NAME main 1312 LOAD_NAME i 1314 CALL_FUNCTION_1 1 '' 1316 CALL_METHOD_1 1 '' 1318 POP_TOP L.1302 1320 LOAD_NAME sleep 1322 LOAD_CONST 3 1324 CALL_FUNCTION_1 1 '' 1326 POP_TOP 1328_1330 JUMP_BACK 1286 'to 1286' 1332_0 COME_FROM 1286 '1286' L.1303 1332 LOAD_NAME os 1334 LOAD_METHOD system 1336 LOAD_NAME os 1338 LOAD_ATTR name 1340 LOAD_STR 'nt' 1342 COMPARE_OP == 1344_1346 POP_JUMP_IF_FALSE 1352 'to 1352' 1348 LOAD_STR 'cls' 1350 JUMP_FORWARD 1354 'to 1354' 1352_0 COME_FROM 1344 '1344' 1352 LOAD_STR 'clear' 1354_0 COME_FROM 1350 '1350' 1354 CALL_METHOD_1 1 '' 1356 POP_TOP L.1304 1358 LOAD_NAME format 1360 LOAD_GLOBAL ERN 1362 LOAD_STR '.8f' 1364 CALL_FUNCTION_2 2 '' 1366 STORE_NAME ER L.1305 1368 LOAD_NAME format 1370 LOAD_GLOBAL WIWCON 1372 LOAD_STR '.8f' 1374 CALL_FUNCTION_2 2 '' 1376 STORE_NAME WC L.1306 1378 LOAD_GLOBAL LICE 1380 LOAD_STR 'VIP' 1382 COMPARE_OP == 1384_1386 POP_JUMP_IF_FALSE 1406 'to 1406' L.1307 1388 LOAD_GLOBAL LICE 1390 FORMAT_VALUE 0 '' 1392 LOAD_STR '(' 1394 LOAD_GLOBAL LIMIT 1396 FORMAT_VALUE 0 '' 1398 LOAD_STR ')' 1400 BUILD_STRING_4 4 1402 STORE_NAME VI 1404 JUMP_FORWARD 1428 'to 1428' 1406_0 COME_FROM 1384 '1384' L.1309 1406 LOAD_NAME Fore 1408 LOAD_ATTR RED 1410 FORMAT_VALUE 0 '' 1412 LOAD_GLOBAL LICE 1414 FORMAT_VALUE 0 '' 1416 LOAD_STR '(' 1418 LOAD_GLOBAL LIMIT 1420 FORMAT_VALUE 0 '' 1422 LOAD_STR ')' 1424 BUILD_STRING_5 5 1426 STORE_NAME VI 1428_0 COME_FROM 1404 '1404' L.1311 1428 LOAD_NAME print 1430 LOAD_NAME banner 1432 CALL_FUNCTION_1 1 '' 1434 POP_TOP L.1312 1436 LOAD_NAME print 1438 LOAD_NAME Fore 1440 LOAD_ATTR GREEN 1442 FORMAT_VALUE 0 '' 1444 LOAD_STR 'Script License ' 1446 LOAD_NAME Fore 1448 LOAD_ATTR RED 1450 FORMAT_VALUE 0 '' 1452 LOAD_STR ':' 1454 LOAD_NAME Fore 1456 LOAD_ATTR RESET 1458 FORMAT_VALUE 0 '' 1460 LOAD_STR ' ' 1462 LOAD_NAME VI 1464 FORMAT_VALUE 0 '' 1466 BUILD_STRING_7 7 1468 CALL_FUNCTION_1 1 '' 1470 POP_TOP L.1313 1472 LOAD_NAME print 1474 LOAD_NAME Fore 1476 LOAD_ATTR GREEN 1478 FORMAT_VALUE 0 '' 1480 LOAD_STR 'Total Loops ' 1482 LOAD_NAME Fore 1484 LOAD_ATTR RED 1486 FORMAT_VALUE 0 '' 1488 LOAD_STR ':' 1490 LOAD_NAME Fore 1492 LOAD_ATTR RESET 1494 FORMAT_VALUE 0 '' 1496 LOAD_STR ' ' 1498 LOAD_NAME LOPS 1500 FORMAT_VALUE 0 '' 1502 BUILD_STRING_7 7 1504 CALL_FUNCTION_1 1 '' 1506 POP_TOP L.1314 1508 LOAD_NAME print 1510 LOAD_NAME Fore 1512 LOAD_ATTR GREEN 1514 FORMAT_VALUE 0 '' 1516 LOAD_STR 'Total accounts ' 1518 LOAD_NAME Fore 1520 LOAD_ATTR RED 1522 FORMAT_VALUE 0 '' 1524 LOAD_STR ':' 1526 LOAD_NAME Fore 1528 LOAD_ATTR RESET 1530 FORMAT_VALUE 0 '' 1532 LOAD_STR ' ' 1534 LOAD_NAME len 1536 LOAD_GLOBAL NUM 1538 CALL_FUNCTION_1 1 '' 1540 FORMAT_VALUE 0 '' 1542 BUILD_STRING_7 7 1544 CALL_FUNCTION_1 1 '' 1546 POP_TOP L.1315 1548 LOAD_NAME print 1550 LOAD_NAME Fore 1552 LOAD_ATTR GREEN 1554 FORMAT_VALUE 0 '' 1556 LOAD_STR 'accounts withdraw ' 1558 LOAD_NAME Fore 1560 LOAD_ATTR RED 1562 FORMAT_VALUE 0 '' 1564 LOAD_STR ':' 1566 LOAD_NAME Fore 1568 LOAD_ATTR RESET 1570 FORMAT_VALUE 0 '' 1572 LOAD_STR ' ' 1574 LOAD_GLOBAL WIWNUM 1576 FORMAT_VALUE 0 '' 1578 LOAD_STR ' account' 1580 BUILD_STRING_8 8 1582 CALL_FUNCTION_1 1 '' 1584 POP_TOP L.1316 1586 LOAD_NAME print 1588 LOAD_NAME Fore 1590 LOAD_ATTR GREEN 1592 FORMAT_VALUE 0 '' 1594 LOAD_STR 'Total withdraw ' 1596 LOAD_NAME Fore 1598 LOAD_ATTR RED 1600 FORMAT_VALUE 0 '' 1602 LOAD_STR ':' 1604 LOAD_NAME Fore 1606 LOAD_ATTR RESET 1608 FORMAT_VALUE 0 '' 1610 LOAD_STR ' ' 1612 LOAD_NAME WC 1614 FORMAT_VALUE 0 '' 1616 LOAD_STR ' ' 1618 LOAD_GLOBAL coin 1620 FORMAT_VALUE 0 '' 1622 BUILD_STRING_9 9 1624 CALL_FUNCTION_1 1 '' 1626 POP_TOP L.1317 1628 LOAD_NAME print 1630 LOAD_NAME Fore 1632 LOAD_ATTR GREEN 1634 FORMAT_VALUE 0 '' 1636 LOAD_STR 'Total balance ' 1638 LOAD_NAME Fore 1640 LOAD_ATTR RED 1642 FORMAT_VALUE 0 '' 1644 LOAD_STR ':' 1646 LOAD_NAME Fore 1648 LOAD_ATTR RESET 1650 FORMAT_VALUE 0 '' 1652 LOAD_STR ' ' 1654 LOAD_NAME ER 1656 FORMAT_VALUE 0 '' 1658 LOAD_STR ' ' 1660 LOAD_GLOBAL coin 1662 FORMAT_VALUE 0 '' 1664 BUILD_STRING_9 9 1666 CALL_FUNCTION_1 1 '' 1668 POP_TOP L.1318 1670 LOAD_NAME print 1672 LOAD_STR '\n' 1674 CALL_FUNCTION_1 1 '' 1676 POP_TOP L.1319 1678 LOAD_GLOBAL SET 1680 LOAD_STR 'AUTOLOOP' 1682 BINARY_SUBSCR 1684 LOAD_METHOD upper 1686 CALL_METHOD_0 0 '' 1688 LOAD_STR 'YES' 1690 COMPARE_OP != 1692_1694 POP_JUMP_IF_FALSE 1702 'to 1702' L.1320 1696 POP_BLOCK 1698_1700 BREAK_LOOP 1760 'to 1760' 1702_0 COME_FROM 1692 '1692' L.1322 1702 LOAD_NAME sleepTime 1704 CALL_FUNCTION_0 0 '' 1706 POP_TOP L.1323 1708 LOAD_CONST 0 1710 STORE_GLOBAL ERN L.1324 1712 LOAD_CONST 0 1714 STORE_NAME INDEX 1716 POP_BLOCK 1718 JUMP_BACK 1254 'to 1254' 1720_0 COME_FROM_FINALLY 1254 '1254' L.1326 1720 DUP_TOP 1722 LOAD_NAME KeyboardInterrupt 1724_1726 <121> 1754 '' 1728 POP_TOP 1730 POP_TOP 1732 POP_TOP L.1327 1734 LOAD_NAME print 1736 LOAD_STR '' 1738 CALL_FUNCTION_1 1 '' 1740 POP_TOP L.1328 1742 LOAD_NAME sys 1744 LOAD_METHOD exit 1746 CALL_METHOD_0 0 '' 1748 POP_TOP 1750 POP_EXCEPT 1752 JUMP_BACK 1254 'to 1254' 1754 <48> 1756_1758 JUMP_BACK 1254 'to 1254' 1760_0 COME_FROM 1698 '1698' Parse error at or near `DUP_TOP' instruction at offset 84 ``` i also attached : .`pyc `file , decompiled results and `pydisasm `results [decompiled_output.zip](https://github.com/rocky/python-decompile3/files/5471265/decompiled_output.zip) [pycfile.zip](https://github.com/rocky/python-decompile3/files/5471266/pycfile.zip) [pydisasm.zip](https://github.com/rocky/python-decompile3/files/5471267/pydisasm.zip)
non_process
parse error at or near dup top python when i decompiled the pyc got the following error messages os win python pydisasm version python bytecode source code size mod instruction context l load name list load genexpr load str make function neither defaults keyword only args annotations nor closures load genexpr load str make function neither defaults keyword only args annotations nor closures load name f in get iter call function get iter call function call function store name requirements pop block load const none dup top dup top call function pop top jump forward to come from with pop jump if true to come from pop top pop top pop top pop except pop top come from file start pyc this code section failed l load const load const none import name re store name re load const load const none import name os store name os load const load const none import name sys store name sys load const load const none import name pkg resources store name pkg resources load const load const none import name random store name random l setup finally to l load name open load str requirements txt call function setup with to store name f in l load name list load genexpr load str make function neither defaults keyword only args annotations nor closures load genexpr load str make function neither defaults keyword only args annotations nor closures load name f in get iter call function get iter call function call function store name requirements pop block load const none dup top dup top call function pop top jump forward to come from with pop jump if true to come from pop top pop top pop top pop except pop top come from l load name requirements get iter come from come from come from for iter to store name require l setup finally to l load name pkg resources load method require load name require call method pop top pop block jump back to come from finally l dup top load name exception pop top store name e pop top setup finally to l load name re load method search load str load name str load name e call function call method load method group load const call method store name a l load name print load str installing load name a binary add load str binary add call function pop top l load name os load method system load str pip install load name a binary add call method pop top pop block pop except load const none store name e delete name e jump back to come from finally load const none store name e delete name e jump back to come from pop block jump forward to come from finally l dup top load name exception pop top store name e pop top setup finally to l load name print load str installation error load name e format value build string call function pop top pop block pop except load const none store name e delete name e jump forward to come from finally load const none store name e delete name e come from come from l load const load const none import name asyncio store name asyncio load const load const none import name logging store name logging load const load const none import name requests store name requests load const load const none import name store name load const load const none import name store name l load const load const sleep strptime import name time import from sleep store name sleep import from strptime store name strptime pop top l load const load const keyboardbuttonurl messagemediacontact replyinlinemarkup updateshortmessage import name attr telethon tl types import from keyboardbuttonurl store name keyboardbuttonurl import from messagemediacontact store name messagemediacontact import from replyinlinemarkup store name replyinlinemarkup import from updateshortmessage store name updateshortmessage pop top l load const load const telegramclient client errors events functions connection import name telethon import from telegramclient store name telegramclient import from client store global client import from errors store name errors import from events store name events import from functions store name functions import from connection store name connection pop top l load const load const joinchannelrequest import name attr telethon tl functions channels import from joinchannelrequest store name joinchannelrequest pop top l load const load const getbotcallbackanswerrequest gethistoryrequest importchatinviterequest getmessagesviewsrequest import name attr telethon tl functions messages import from getbotcallbackanswerrequest store name getbotcallbackanswerrequest import from gethistoryrequest store name gethistoryrequest import from importchatinviterequest store name importchatinviterequest import from getmessagesviewsrequest store name getmessagesviewsrequest pop top l load const load const datetime timedelta import name datetime import from datetime store name datetime import from timedelta store name timedelta pop top l load const load const fore style init import name colorama import from fore store name fore import from style store name style import from init store name color ama pop top l load const load const beautifulsoup import name import from beautifulsoup store name beautifulsoup pop top l load const load const none import name random store name random l load const load const none import name socks store name socks l load name logging load attr basicconfig load name logging load attr error load const level call function kw total positional and keyword args pop top l load name color ama load const true load const autoreset call function kw total positional and keyword args pop top l load name os load method system load name os load attr name load str nt compare op pop jump if false to load str cls jump forward to come from load str clear come from call method pop top l load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name fore load attr blue l format value load str edit by l load name style load attr dim load name fore load attr red binary add l format value load str l load name fore load attr yellow l format value load str abbas bachari l load name fore load attr blue l format value load str version l load name fore load attr yellow l format value load str n l load name style load attr normal load name fore load attr red binary add l format value load str n l load name style load attr bright load name fore load attr green binary add l format value load str sponsor channel l load name style load attr dim load name fore load attr red binary add l format value load str l load name style load attr reset all l format value load str build string store name banner l load const load str bitcoinclick bot load str start load const id username start build const key map l load const load str litecoin click bot load str start azyg load const id username start build const key map l load const load str dogecoin click bot load str start ljsp load const id username start build const key map l load const load str bch clickbot load str start load const id username start build const key map l load const load str zcash click bot load str start load const id username start build const key map l load const btc ltc doge bch zec build const key map store name symbols l load const store global api id l load str store global api hash l load str store name msg l load name requests load method session call method store global session l load str user agent load str mozilla linux android moto g play applewebkit khtml like gecko chrome mobile safari build map store name ua l load str store global dogeclick channel l load str store global phone number l load str store global wallet l load str store global isvisit l load str store global isjoin l load str store global isbotms l load str store global isloop l load str store global balance l build map store global set l load str store global coin l load const store name index l load const store global num l load str sit store global stat l load const none store global captcha l load const none store global sit l load str store global message l load str store global channel name l load const none store global channel id l load str store global bot name l load str store global coment l load const store name apver l build list store global bots l build list store global chns l build list store global grps l build list store global usrs l load const store global limit l load str store global lice l load str store global mase l load const store global ver l load const store global mass id l load const none store global click data l load str store global vipmes l load const store global jnum l load const store global ern l load const store global wiwnum l load const store global wiwcon l load const store name lops l build map store global setin l build list store global plogins l build map store global mesegs l build map store global admeseg l load str yes store global add l load const none store global minwd l load str numbers txt store global path l load const none store global proxy l load code load str isnumber make function neither defaults keyword only args annotations nor closures store name isnumber l load code load str request make function neither defaults keyword only args annotations nor closures store name request l load code load str getip make function neither defaults keyword only args annotations nor closures store name getip l load code load str send contact make function neither defaults keyword only args annotations nor closures store name send contact l load code load str send messege to members make function neither defaults keyword only args annotations nor closures store name send messege to members l load code load str chnum make function neither defaults keyword only args annotations nor closures store name chnum l load code load str numbers make function neither defaults keyword only args annotations nor closures store name numbers l load code load str anonsrobot make function neither defaults keyword only args annotations nor closures store name anonsrobot l load code load str start plogins make function neither defaults keyword only args annotations nor closures store name start plogins l load code load str send plogin messege make function neither defaults keyword only args annotations nor closures store name send plogin messege l load code load str send group messege make function neither defaults keyword only args annotations nor closures store name send group messege l load code load str settings make function neither defaults keyword only args annotations nor closures store name settings l load code load str send to privait group make function neither defaults keyword only args annotations nor closures store name send to privait group l load code load str creat user make function neither defaults keyword only args annotations nor closures store name creat user l load code load str add plogin make function neither defaults keyword only args annotations nor closures store name add plogin l load code load str delete plogin make function neither defaults keyword only args annotations nor closures store name delete plogin l load code load str getloop make function neither defaults keyword only args annotations nor closures store name getloop l load code load str add channel make function neither defaults keyword only args annotations nor closures store name add channel l load str no store global iswid l load code load str delete cannels make function neither defaults keyword only args annotations nor closures store name delete cannels l load code load str reset make function neither defaults keyword only args annotations nor closures store name reset l load code load str maseg make function neither defaults keyword only args annotations nor closures store name maseg l load code load str wait make function neither defaults keyword only args annotations nor closures store name wait l load code load str floodtimer make function neither defaults keyword only args annotations nor closures store name floodtimer l load code load str sleeptime make function neither defaults keyword only args annotations nor closures store name sleeptime l load code load str jointimer make function neither defaults keyword only args annotations nor closures store name jointimer l load code load str get username make function neither defaults keyword only args annotations nor closures store name get username l load code load str forward messages make function neither defaults keyword only args annotations nor closures store name forward messages l load code load str start make function neither defaults keyword only args annotations nor closures store name start l load code load str main make function neither defaults keyword only args annotations nor closures store name main come from come from come from l setup finally to l load name asyncio load method get event loop call method load method run until complete load name numbers call function call method pop top l load name lops load const inplace add store name lops l load global num get iter come from for iter to store name i l load name index load const inplace add store name index l load const store global jnum l load name asyncio load method get event loop call method load method run until complete load name main load name i call function call method pop top l load name sleep load const call function pop top jump back to come from l load name os load method system load name os load attr name load str nt compare op pop jump if false to load str cls jump forward to come from load str clear come from call method pop top l load name format load global ern load str call function store name er l load name format load global wiwcon load str call function store name wc l load global lice load str vip compare op pop jump if false to l load global lice format value load str load global limit format value load str build string store name vi jump forward to come from l load name fore load attr red format value load global lice format value load str load global limit format value load str build string store name vi come from l load name print load name banner call function pop top l load name print load name fore load attr green format value load str script license load name fore load attr red format value load str load name fore load attr reset format value load str load name vi format value build string call function pop top l load name print load name fore load attr green format value load str total loops load name fore load attr red format value load str load name fore load attr reset format value load str load name lops format value build string call function pop top l load name print load name fore load attr green format value load str total accounts load name fore load attr red format value load str load name fore load attr reset format value load str load name len load global num call function format value build string call function pop top l load name print load name fore load attr green format value load str accounts withdraw load name fore load attr red format value load str load name fore load attr reset format value load str load global wiwnum format value load str account build string call function pop top l load name print load name fore load attr green format value load str total withdraw load name fore load attr red format value load str load name fore load attr reset format value load str load name wc format value load str load global coin format value build string call function pop top l load name print load name fore load attr green format value load str total balance load name fore load attr red format value load str load name fore load attr reset format value load str load name er format value load str load global coin format value build string call function pop top l load name print load str n call function pop top l load global set load str autoloop binary subscr load method upper call method load str yes compare op pop jump if false to l pop block break loop to come from l load name sleeptime call function pop top l load const store global ern l load const store name index pop block jump back to come from finally l dup top load name keyboardinterrupt pop top pop top pop top l load name print load str call function pop top l load name sys load method exit call method pop top pop except jump back to jump back to come from parse error at or near dup top instruction at offset i also attached pyc file decompiled results and pydisasm results
0
6,211
9,124,270,989
IssuesEvent
2019-02-24 01:15:30
ahsankhatri/firestore-php
https://api.github.com/repos/ahsankhatri/firestore-php
closed
Error: PERMISSION_DENIED: Missing or insufficient permissions
enhancement in-process
I tried to write to FireStore but the server responded 403 PERMISSION_DENIED: Missing or insufficient permissions. The database rule is: ``` service cloud.firestore { match /databases/{database}/documents { match /{document=**} { allow read, write: if request.auth.uid != null; } } } ``` If I change rule to `allow read, write: if true;` (turn off database security), it's OK.
1.0
Error: PERMISSION_DENIED: Missing or insufficient permissions - I tried to write to FireStore but the server responded 403 PERMISSION_DENIED: Missing or insufficient permissions. The database rule is: ``` service cloud.firestore { match /databases/{database}/documents { match /{document=**} { allow read, write: if request.auth.uid != null; } } } ``` If I change rule to `allow read, write: if true;` (turn off database security), it's OK.
process
error permission denied missing or insufficient permissions i tried to write to firestore but the server responded permission denied missing or insufficient permissions the database rule is service cloud firestore match databases database documents match document allow read write if request auth uid null if i change rule to allow read write if true turn off database security it s ok
1
1,548
4,155,379,708
IssuesEvent
2016-06-16 14:45:58
openvstorage/openvstorage-health-check
https://api.github.com/repos/openvstorage/openvstorage-health-check
closed
Vpool is not present, ignore the test on that node
priority_urgent process_duplicate type_bug
If no vpool is present on a compute, controller ignore the test for that specific node.
1.0
Vpool is not present, ignore the test on that node - If no vpool is present on a compute, controller ignore the test for that specific node.
process
vpool is not present ignore the test on that node if no vpool is present on a compute controller ignore the test for that specific node
1
9,714
25,103,837,955
IssuesEvent
2022-11-08 15:19:49
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
Invalid C# syntax in code sample
doc-bug assigned-to-author triaged architecture-center/svc example-scenario/subsvc Pri2
Hello, It appears that the code sample for the referenced documentation contains some invalid C# syntax, and given the code, it's not clear how to check if the device needs an SMS wakeup. `if (var == [DeviceNotConnected] && connTimeOut == 0)` I'm not sure where "DeviceNotConnected" is coming from, and I don't believe the use of `var` here or later in the method is valid. Full code sample: `TimeSpan connTimeOut = FromSeconds(0); // Period to wait for device to connect. TimeSpan funcTimeOut = FromSeconds(30); // Period to wait for method to execute. while (true) { // Send the command via direct method. Initially use a timeout of zero // for the connection, which determines whether the device is connected to // IoT Hub or needs an SMS wakeup sent to it. var method = new CloudToDeviceMethod("RemoteCommand", funcTimeOut, connTimeOut); methodInvocation1.SetPayloadJson(CommandPayload); var response = await serviceClient.InvokeDeviceMethodAsync(deviceId, method); if (var == [DeviceNotConnected] && connTimeOut == 0) { // The device is not currently connected and needs an SMS wakeup. This // device should wake up within a period of < 30 seconds. Send the wakeup // and retry the method request with a 30 second timeout on waiting for // the device to connect. connTimeOut = FromSeconds(30); // Set a 30 second connection timeout. SendAsyncSMSWakeUpToDevice(); // Send SMS wakeup through mobile gateway. continue; // Retry with new connection timeout. } else { // The method either succeeded or failed. ActOnMethodResult(var); break; } }` Thank you! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7b2f7457-51f0-b9e0-04bd-dabbe3586b34 * Version Independent ID: 08b9268e-c8d1-5259-0f0d-3fc851f35dbe * Content: [IoT application-to-device commands - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/iot/cloud-to-device) * Content Source: [docs/example-scenario/iot/cloud-to-device.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/iot/cloud-to-device.yml) * Service: **architecture-center** * Sub-service: **example-scenario** * GitHub Login: @wamachine * Microsoft Alias: **pnp**
1.0
Invalid C# syntax in code sample - Hello, It appears that the code sample for the referenced documentation contains some invalid C# syntax, and given the code, it's not clear how to check if the device needs an SMS wakeup. `if (var == [DeviceNotConnected] && connTimeOut == 0)` I'm not sure where "DeviceNotConnected" is coming from, and I don't believe the use of `var` here or later in the method is valid. Full code sample: `TimeSpan connTimeOut = FromSeconds(0); // Period to wait for device to connect. TimeSpan funcTimeOut = FromSeconds(30); // Period to wait for method to execute. while (true) { // Send the command via direct method. Initially use a timeout of zero // for the connection, which determines whether the device is connected to // IoT Hub or needs an SMS wakeup sent to it. var method = new CloudToDeviceMethod("RemoteCommand", funcTimeOut, connTimeOut); methodInvocation1.SetPayloadJson(CommandPayload); var response = await serviceClient.InvokeDeviceMethodAsync(deviceId, method); if (var == [DeviceNotConnected] && connTimeOut == 0) { // The device is not currently connected and needs an SMS wakeup. This // device should wake up within a period of < 30 seconds. Send the wakeup // and retry the method request with a 30 second timeout on waiting for // the device to connect. connTimeOut = FromSeconds(30); // Set a 30 second connection timeout. SendAsyncSMSWakeUpToDevice(); // Send SMS wakeup through mobile gateway. continue; // Retry with new connection timeout. } else { // The method either succeeded or failed. ActOnMethodResult(var); break; } }` Thank you! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7b2f7457-51f0-b9e0-04bd-dabbe3586b34 * Version Independent ID: 08b9268e-c8d1-5259-0f0d-3fc851f35dbe * Content: [IoT application-to-device commands - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/iot/cloud-to-device) * Content Source: [docs/example-scenario/iot/cloud-to-device.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/iot/cloud-to-device.yml) * Service: **architecture-center** * Sub-service: **example-scenario** * GitHub Login: @wamachine * Microsoft Alias: **pnp**
non_process
invalid c syntax in code sample hello it appears that the code sample for the referenced documentation contains some invalid c syntax and given the code it s not clear how to check if the device needs an sms wakeup if var conntimeout i m not sure where devicenotconnected is coming from and i don t believe the use of var here or later in the method is valid full code sample timespan conntimeout fromseconds period to wait for device to connect timespan functimeout fromseconds period to wait for method to execute while true send the command via direct method initially use a timeout of zero for the connection which determines whether the device is connected to iot hub or needs an sms wakeup sent to it var method new cloudtodevicemethod remotecommand functimeout conntimeout setpayloadjson commandpayload var response await serviceclient invokedevicemethodasync deviceid method if var conntimeout the device is not currently connected and needs an sms wakeup this device should wake up within a period of seconds send the wakeup and retry the method request with a second timeout on waiting for the device to connect conntimeout fromseconds set a second connection timeout sendasyncsmswakeuptodevice send sms wakeup through mobile gateway continue retry with new connection timeout else the method either succeeded or failed actonmethodresult var break thank you document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service example scenario github login wamachine microsoft alias pnp
0
39,335
12,663,382,544
IssuesEvent
2020-06-18 01:10:40
kenferrara/atlasdb
https://api.github.com/repos/kenferrara/atlasdb
opened
CVE-2020-7238 (High) detected in netty-codec-http-4.1.39.Final.jar, netty-all-4.1.39.Final.jar
security vulnerability
## CVE-2020-7238 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>netty-codec-http-4.1.39.Final.jar</b>, <b>netty-all-4.1.39.Final.jar</b></p></summary> <p> <details><summary><b>netty-codec-http-4.1.39.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p> <p> Dependency Hierarchy: - netty-handler-proxy-4.1.39.Final.jar (Root Library) - :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library) </details> <details><summary><b>netty-all-4.1.39.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra-integration-tests/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar</p> <p> Dependency Hierarchy: - cassandra-all-2.2.13-1.6.0.jar (Root Library) - :x: **netty-all-4.1.39.Final.jar** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869. <p>Publish Date: 2020-01-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238>CVE-2020-7238</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/netty/netty/issues/9861">https://github.com/netty/netty/issues/9861</a></p> <p>Release Date: 2020-01-27</p> <p>Fix Resolution: io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"io.netty:netty-handler-proxy:4.1.39.Final;io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"},{"packageType":"Java","groupId":"io.netty","packageName":"netty-all","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"com.palantir.cassandra:cassandra-all:2.2.13-1.6.0;io.netty:netty-all:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"}],"vulnerabilityIdentifier":"CVE-2020-7238","vulnerabilityDetails":"Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7238 (High) detected in netty-codec-http-4.1.39.Final.jar, netty-all-4.1.39.Final.jar - ## CVE-2020-7238 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>netty-codec-http-4.1.39.Final.jar</b>, <b>netty-all-4.1.39.Final.jar</b></p></summary> <p> <details><summary><b>netty-codec-http-4.1.39.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p> <p> Dependency Hierarchy: - netty-handler-proxy-4.1.39.Final.jar (Root Library) - :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library) </details> <details><summary><b>netty-all-4.1.39.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra-integration-tests/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar</p> <p> Dependency Hierarchy: - cassandra-all-2.2.13-1.6.0.jar (Root Library) - :x: **netty-all-4.1.39.Final.jar** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869. <p>Publish Date: 2020-01-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238>CVE-2020-7238</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/netty/netty/issues/9861">https://github.com/netty/netty/issues/9861</a></p> <p>Release Date: 2020-01-27</p> <p>Fix Resolution: io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"io.netty:netty-handler-proxy:4.1.39.Final;io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"},{"packageType":"Java","groupId":"io.netty","packageName":"netty-all","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"com.palantir.cassandra:cassandra-all:2.2.13-1.6.0;io.netty:netty-all:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"}],"vulnerabilityIdentifier":"CVE-2020-7238","vulnerabilityDetails":"Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in netty codec http final jar netty all final jar cve high severity vulnerability vulnerable libraries netty codec http final jar netty all final jar netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file tmp ws scm atlasdb atlasdb cassandra build gradle path to vulnerable library root gradle caches modules files io netty netty codec http final netty codec http final jar root gradle caches modules files io netty netty codec http final netty codec http final jar dependency hierarchy netty handler proxy final jar root library x netty codec http final jar vulnerable library netty all final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file tmp ws scm atlasdb atlasdb cassandra integration tests build gradle path to vulnerable library root gradle caches modules files io netty netty all final netty all final jar root gradle caches modules files io netty netty all final netty all final jar dependency hierarchy cassandra all jar root library x netty all final jar vulnerable library vulnerability details netty final allows http request smuggling because it mishandles transfer encoding whitespace such as a transfer encoding chunked line and a later content length header this issue exists because of an incomplete fix for cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all final io netty netty codec http final isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails netty final allows http request smuggling because it mishandles transfer encoding whitespace such as a transfer encoding chunked line and a later content length header this issue exists because of an incomplete fix for cve vulnerabilityurl
0
207,389
7,127,842,932
IssuesEvent
2018-01-21 03:16:40
DASSL/ClassDB
https://api.github.com/repos/DASSL/ClassDB
closed
The docs-development process may not support adding image and other files (WM)
missing priority low wrong
**PS:** I am creating this issue so we can collectively discuss and receive notifications as new comments are added. Context: We have add at least two instances where non-markdown files are to be included in the docs folder. See #136 and #139. In general, we have the need to add HTML/Visio/image files to the docs folder with design notes and these files will have to be linked to from docs pages which are markdown (MD) files. On the surface, this does not sound like much of an issue, but I worry our development process for docs may not support this ability: at least not very easily. Our docs-development process uses the Wiki as the "development environment" to write the docs and it worked very well in M1: We need to use the Wiki so anyone can easily update any page directly online and also avoid the approval process which can be both labor and time intensive. It seems, if we want to continue to use the Wiki for doc development, we will need to clone the Wiki as a repository, add files to it, and check in the file. Even then, I am not sure if it will work correctly as evidenced in [@afig's Wiki page](https://github.com/DASSL/ClassDB/wiki/createRole()-User-Schema-checks-and-actions) on the flowchart related to Issue #136. See how the embedded image referencing a remote URL does not show up. We need to **urgently** understand the underlying issues and learn workarounds/solutions so documentation does not become a problem. Perhaps there is no issue at all, in which case, it will be helpful to confirm so before we get too far and then find a problem.
1.0
The docs-development process may not support adding image and other files (WM) - **PS:** I am creating this issue so we can collectively discuss and receive notifications as new comments are added. Context: We have add at least two instances where non-markdown files are to be included in the docs folder. See #136 and #139. In general, we have the need to add HTML/Visio/image files to the docs folder with design notes and these files will have to be linked to from docs pages which are markdown (MD) files. On the surface, this does not sound like much of an issue, but I worry our development process for docs may not support this ability: at least not very easily. Our docs-development process uses the Wiki as the "development environment" to write the docs and it worked very well in M1: We need to use the Wiki so anyone can easily update any page directly online and also avoid the approval process which can be both labor and time intensive. It seems, if we want to continue to use the Wiki for doc development, we will need to clone the Wiki as a repository, add files to it, and check in the file. Even then, I am not sure if it will work correctly as evidenced in [@afig's Wiki page](https://github.com/DASSL/ClassDB/wiki/createRole()-User-Schema-checks-and-actions) on the flowchart related to Issue #136. See how the embedded image referencing a remote URL does not show up. We need to **urgently** understand the underlying issues and learn workarounds/solutions so documentation does not become a problem. Perhaps there is no issue at all, in which case, it will be helpful to confirm so before we get too far and then find a problem.
non_process
the docs development process may not support adding image and other files wm ps i am creating this issue so we can collectively discuss and receive notifications as new comments are added context we have add at least two instances where non markdown files are to be included in the docs folder see and in general we have the need to add html visio image files to the docs folder with design notes and these files will have to be linked to from docs pages which are markdown md files on the surface this does not sound like much of an issue but i worry our development process for docs may not support this ability at least not very easily our docs development process uses the wiki as the development environment to write the docs and it worked very well in we need to use the wiki so anyone can easily update any page directly online and also avoid the approval process which can be both labor and time intensive it seems if we want to continue to use the wiki for doc development we will need to clone the wiki as a repository add files to it and check in the file even then i am not sure if it will work correctly as evidenced in on the flowchart related to issue see how the embedded image referencing a remote url does not show up we need to urgently understand the underlying issues and learn workarounds solutions so documentation does not become a problem perhaps there is no issue at all in which case it will be helpful to confirm so before we get too far and then find a problem
0
8,279
11,434,738,769
IssuesEvent
2020-02-04 17:58:25
prisma/prisma-client-js
https://api.github.com/repos/prisma/prisma-client-js
closed
Update graphics to use schema.prisma instead of datamodel.mdl
process/candidate
## What Currently the graphics use `datamodel.mdl`, e.g. <img width="751" alt="Screenshot 2020-02-04 at 15 51 58" src="https://user-images.githubusercontent.com/1992255/73755537-7269e400-4766-11ea-8bb7-420b4a5b0135.png"> They should be updated to `schema.prisma`
1.0
Update graphics to use schema.prisma instead of datamodel.mdl - ## What Currently the graphics use `datamodel.mdl`, e.g. <img width="751" alt="Screenshot 2020-02-04 at 15 51 58" src="https://user-images.githubusercontent.com/1992255/73755537-7269e400-4766-11ea-8bb7-420b4a5b0135.png"> They should be updated to `schema.prisma`
process
update graphics to use schema prisma instead of datamodel mdl what currently the graphics use datamodel mdl e g img width alt screenshot at src they should be updated to schema prisma
1
1,601
3,070,690,084
IssuesEvent
2015-08-19 07:06:20
NatLibFi/Skosmos
https://api.github.com/repos/NatLibFi/Skosmos
opened
Make use of Jena3 text index for better performance
branch enhancement performance size-large
There are significant changes (implemented by Alexis Miara and myself) in the [jena-text](http://jena.staging.apache.org/documentation/query/text-query.html) module of Jena 3.0.0 / Fuseki 1.3.0 / Fuseki 2.3.0. These include * support for storing language tags of literals and limiting queries to a specific language * support for storing full literal values in the index and accessing them at query time * support for deleting obsolete entries from the text index These together enable a new way of using the text index from Skosmos: * Text queries could, in most cases, be limited to a specific language. This avoids false hits from the text index that would have to be filtered out using SPARQL, and should thus speed up queries, particularly the alphabetical display for large vocabularies. * Since the text index can return full literal values, there is less need to find out which literal value actually matched the query (using regular expressions or string matching functions, as is done currently). This should make text index related SPARQL queries both simpler and faster. * The `uidField` should be enabled, so that stale entries will be dropped from the text index. Currently the performance of text index related queries deteriorates slightly each time the vocabulary data is updated. This is probably due to stale entries. Cleaning them up should prevent this performance deterioration. Text index related code in JenaTextSparql (and possibly GenericSparql) will need to be heavily rewritten. Luckily the new code should be simpler than the old one and we already have pretty good unit tests for this functionality, so it is easy to verify what works and what doesn't. Text index configuration needs to be changed to enable the new features, and text indexes must then be rebuilt. Fuseki 1.3.0/2.3.0 will require Java 8 to be installed on servers, development machines and the Travis CI environment (where it should be available, but not used by default). (Finto project note: this is a way of implementing FINTO-85: Tuki hyvin suurille tietovarannoille)
True
Make use of Jena3 text index for better performance - There are significant changes (implemented by Alexis Miara and myself) in the [jena-text](http://jena.staging.apache.org/documentation/query/text-query.html) module of Jena 3.0.0 / Fuseki 1.3.0 / Fuseki 2.3.0. These include * support for storing language tags of literals and limiting queries to a specific language * support for storing full literal values in the index and accessing them at query time * support for deleting obsolete entries from the text index These together enable a new way of using the text index from Skosmos: * Text queries could, in most cases, be limited to a specific language. This avoids false hits from the text index that would have to be filtered out using SPARQL, and should thus speed up queries, particularly the alphabetical display for large vocabularies. * Since the text index can return full literal values, there is less need to find out which literal value actually matched the query (using regular expressions or string matching functions, as is done currently). This should make text index related SPARQL queries both simpler and faster. * The `uidField` should be enabled, so that stale entries will be dropped from the text index. Currently the performance of text index related queries deteriorates slightly each time the vocabulary data is updated. This is probably due to stale entries. Cleaning them up should prevent this performance deterioration. Text index related code in JenaTextSparql (and possibly GenericSparql) will need to be heavily rewritten. Luckily the new code should be simpler than the old one and we already have pretty good unit tests for this functionality, so it is easy to verify what works and what doesn't. Text index configuration needs to be changed to enable the new features, and text indexes must then be rebuilt. Fuseki 1.3.0/2.3.0 will require Java 8 to be installed on servers, development machines and the Travis CI environment (where it should be available, but not used by default). (Finto project note: this is a way of implementing FINTO-85: Tuki hyvin suurille tietovarannoille)
non_process
make use of text index for better performance there are significant changes implemented by alexis miara and myself in the module of jena fuseki fuseki these include support for storing language tags of literals and limiting queries to a specific language support for storing full literal values in the index and accessing them at query time support for deleting obsolete entries from the text index these together enable a new way of using the text index from skosmos text queries could in most cases be limited to a specific language this avoids false hits from the text index that would have to be filtered out using sparql and should thus speed up queries particularly the alphabetical display for large vocabularies since the text index can return full literal values there is less need to find out which literal value actually matched the query using regular expressions or string matching functions as is done currently this should make text index related sparql queries both simpler and faster the uidfield should be enabled so that stale entries will be dropped from the text index currently the performance of text index related queries deteriorates slightly each time the vocabulary data is updated this is probably due to stale entries cleaning them up should prevent this performance deterioration text index related code in jenatextsparql and possibly genericsparql will need to be heavily rewritten luckily the new code should be simpler than the old one and we already have pretty good unit tests for this functionality so it is easy to verify what works and what doesn t text index configuration needs to be changed to enable the new features and text indexes must then be rebuilt fuseki will require java to be installed on servers development machines and the travis ci environment where it should be available but not used by default finto project note this is a way of implementing finto tuki hyvin suurille tietovarannoille
0
12,127
2,685,016,856
IssuesEvent
2015-03-29 16:24:00
IssueMigrationTest/Test5
https://api.github.com/repos/IssueMigrationTest/Test5
closed
*ERROR* basedados.py:11: cannot locate module: psycopg2
auto-migrated Priority-Medium Type-Defect
**Issue by luisl...@gmail.com** _9 Jul 2009 at 2:09 GMT_ _Originally opened on Google Code_ ---- ``` Hi, I have a module (basedados.py) that starts as: """ #-*- coding: iso-8859-15 -*- import ConfigParser import psycopg2 """ It runs fine in Python, as stand-alone or when called from other modules. If I try: $ shedskin basedados *** SHED SKIN Python-to-C++ Compiler 0.1.1 *** Copyright 2005-2009 Mark Dufour; License GNU GPL version 3 (See LICENSE) *ERROR* basedados.py:11: cannot locate module: psycopg2 Even if I remove every line from the module except the 'import psycopg2' statement, the error is the same. I'm using Slackware Linux 12.1, python 2.5.2 and shedskin 0.1.1 ```
1.0
*ERROR* basedados.py:11: cannot locate module: psycopg2 - **Issue by luisl...@gmail.com** _9 Jul 2009 at 2:09 GMT_ _Originally opened on Google Code_ ---- ``` Hi, I have a module (basedados.py) that starts as: """ #-*- coding: iso-8859-15 -*- import ConfigParser import psycopg2 """ It runs fine in Python, as stand-alone or when called from other modules. If I try: $ shedskin basedados *** SHED SKIN Python-to-C++ Compiler 0.1.1 *** Copyright 2005-2009 Mark Dufour; License GNU GPL version 3 (See LICENSE) *ERROR* basedados.py:11: cannot locate module: psycopg2 Even if I remove every line from the module except the 'import psycopg2' statement, the error is the same. I'm using Slackware Linux 12.1, python 2.5.2 and shedskin 0.1.1 ```
non_process
error basedados py cannot locate module issue by luisl gmail com jul at gmt originally opened on google code hi i have a module basedados py that starts as coding iso import configparser import it runs fine in python as stand alone or when called from other modules if i try shedskin basedados shed skin python to c compiler copyright mark dufour license gnu gpl version see license error basedados py cannot locate module even if i remove every line from the module except the import statement the error is the same i m using slackware linux python and shedskin
0
14,936
18,365,441,937
IssuesEvent
2021-10-10 00:48:00
FTBTeam/FTB-App-Feedback
https://api.github.com/repos/FTBTeam/FTB-App-Feedback
closed
[Bug] Some files still are stored in default install location (/users/my_username/.ftba) even when install location is changed
bug bug:subprocess effect:linux bug:unconfirmed subprocess
**Operating System:** OS: Windows **Installer Version:** Installer: **App Version:** App: 202007161516-479dedc615 **UI Version:** UI: 39f37b8d **Describe the bug** Even when I change in install location in the launcher, some necessary files are installed in the default location (in the bin folder). It seems these files are used for the vanilla launcher. If I delete the folder, the launcher no longer can run instances. If I move the files from there to my chosen install location, the launcher still no longer can run instances. **To Reproduce** Steps to reproduce the behaviour: 1. Change install location 2. create a new instance 3. run the instance once 4. Observe existence of .ftba folder in home directory with necessary files still located there **Expected behaviour** All files necessary to the launcher are installed in the chosen install location **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here.
2.0
[Bug] Some files still are stored in default install location (/users/my_username/.ftba) even when install location is changed - **Operating System:** OS: Windows **Installer Version:** Installer: **App Version:** App: 202007161516-479dedc615 **UI Version:** UI: 39f37b8d **Describe the bug** Even when I change in install location in the launcher, some necessary files are installed in the default location (in the bin folder). It seems these files are used for the vanilla launcher. If I delete the folder, the launcher no longer can run instances. If I move the files from there to my chosen install location, the launcher still no longer can run instances. **To Reproduce** Steps to reproduce the behaviour: 1. Change install location 2. create a new instance 3. run the instance once 4. Observe existence of .ftba folder in home directory with necessary files still located there **Expected behaviour** All files necessary to the launcher are installed in the chosen install location **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here.
process
some files still are stored in default install location users my username ftba even when install location is changed operating system os windows installer version installer app version app ui version ui describe the bug even when i change in install location in the launcher some necessary files are installed in the default location in the bin folder it seems these files are used for the vanilla launcher if i delete the folder the launcher no longer can run instances if i move the files from there to my chosen install location the launcher still no longer can run instances to reproduce steps to reproduce the behaviour change install location create a new instance run the instance once observe existence of ftba folder in home directory with necessary files still located there expected behaviour all files necessary to the launcher are installed in the chosen install location screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here
1
10,473
12,421,425,777
IssuesEvent
2020-05-23 16:48:11
facebook/hhvm
https://api.github.com/repos/facebook/hhvm
closed
Unable append to array returned by function as referece
php5 incompatibility probably easy
In this case (http://3v4l.org/DmV7V) appending to array fails (unlike PHP interpreter) ``` php <?php class Test { private $array = []; public function __construct() { $i = 0; while ($i < 10) { $this->array[$i] = str_repeat(" ", 1e6); $i++; } } public function &variables() { return $this->array; } } $test = new Test(); $test->variables()[10] = str_repeat(" ", 1e6); // ok $test->variables()[] = str_repeat(" ", 1e6); // error, why ? echo count($test->variables()) . "\n"; ```
True
Unable append to array returned by function as referece - In this case (http://3v4l.org/DmV7V) appending to array fails (unlike PHP interpreter) ``` php <?php class Test { private $array = []; public function __construct() { $i = 0; while ($i < 10) { $this->array[$i] = str_repeat(" ", 1e6); $i++; } } public function &variables() { return $this->array; } } $test = new Test(); $test->variables()[10] = str_repeat(" ", 1e6); // ok $test->variables()[] = str_repeat(" ", 1e6); // error, why ? echo count($test->variables()) . "\n"; ```
non_process
unable append to array returned by function as referece in this case appending to array fails unlike php interpreter php php class test private array public function construct i while i this array str repeat i public function variables return this array test new test test variables str repeat ok test variables str repeat error why echo count test variables n
0
500,555
14,502,050,319
IssuesEvent
2020-12-11 20:25:38
GetSpeckle/speckle-browser-extension
https://api.github.com/repos/GetSpeckle/speckle-browser-extension
closed
Upgrade polkadot libs to support edgeware nodes
Priority: Medium
While speckle is still using `@polkadot/api 0.76.1` and `@polkadot/keyring 0.76.1`, `@polkadot/api 0.81.1` and `@polkadot/keyring 0.93.1` are already released as of writing. `edgeware-node-types` relies on those new libs, hence speckle cannot support edgeware until we upgrade to those versions. Certain interfaces are changed in those versions, `keyring-vault` needs to be updated accordingly
1.0
Upgrade polkadot libs to support edgeware nodes - While speckle is still using `@polkadot/api 0.76.1` and `@polkadot/keyring 0.76.1`, `@polkadot/api 0.81.1` and `@polkadot/keyring 0.93.1` are already released as of writing. `edgeware-node-types` relies on those new libs, hence speckle cannot support edgeware until we upgrade to those versions. Certain interfaces are changed in those versions, `keyring-vault` needs to be updated accordingly
non_process
upgrade polkadot libs to support edgeware nodes while speckle is still using polkadot api and polkadot keyring polkadot api and polkadot keyring are already released as of writing edgeware node types relies on those new libs hence speckle cannot support edgeware until we upgrade to those versions certain interfaces are changed in those versions keyring vault needs to be updated accordingly
0
13,911
16,668,775,354
IssuesEvent
2021-06-07 08:19:52
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
opened
Release 4.2 - June 2021
P1 release team-XProduct type: process
# Status of Bazel 4.2 This release will use Bazel 4.1.0 as its baseline and we will apply selected cherry-picks and backports on top of it. Please request cherry-picks that you'd like to get into Bazel 4.2.0 here via a comment. - Expected release date: June 2021 - [List of release blockers](https://github.com/bazelbuild/bazel/labels/Release%20blocker) To report a release-blocking bug, please file a bug using the `Release blocker` label, and cc me. Task list: - [ ] Pick release baseline: https://github.com/bazelbuild/bazel/commit/c4975efdd7ede7b46637bf353209d9ac371181a5 - [ ] Create release candidate: https://releases.bazel.build/4.2.0/rc1/ - [ ] Check downstream projects: ... - [ ] [Create draft release announcement](https://docs.google.com/document/d/1wDvulLlj4NAlPZamdlEVFORks3YXJonCjyuQMUQEmB0/edit) - [ ] Send for review the release announcement PR: ... - [ ] Push the release, notify package maintainers: ... - [ ] Update the documentation - [ ] Push the blog post - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/tag/4.2.0)
1.0
Release 4.2 - June 2021 - # Status of Bazel 4.2 This release will use Bazel 4.1.0 as its baseline and we will apply selected cherry-picks and backports on top of it. Please request cherry-picks that you'd like to get into Bazel 4.2.0 here via a comment. - Expected release date: June 2021 - [List of release blockers](https://github.com/bazelbuild/bazel/labels/Release%20blocker) To report a release-blocking bug, please file a bug using the `Release blocker` label, and cc me. Task list: - [ ] Pick release baseline: https://github.com/bazelbuild/bazel/commit/c4975efdd7ede7b46637bf353209d9ac371181a5 - [ ] Create release candidate: https://releases.bazel.build/4.2.0/rc1/ - [ ] Check downstream projects: ... - [ ] [Create draft release announcement](https://docs.google.com/document/d/1wDvulLlj4NAlPZamdlEVFORks3YXJonCjyuQMUQEmB0/edit) - [ ] Send for review the release announcement PR: ... - [ ] Push the release, notify package maintainers: ... - [ ] Update the documentation - [ ] Push the blog post - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/tag/4.2.0)
process
release june status of bazel this release will use bazel as its baseline and we will apply selected cherry picks and backports on top of it please request cherry picks that you d like to get into bazel here via a comment expected release date june to report a release blocking bug please file a bug using the release blocker label and cc me task list pick release baseline create release candidate check downstream projects send for review the release announcement pr push the release notify package maintainers update the documentation push the blog post update the
1
17,729
23,636,906,817
IssuesEvent
2022-08-25 13:58:01
argosp/trialdash
https://api.github.com/repos/argosp/trialdash
closed
Position in one shape entities from different types
Priority High in process
in the current staged app there is no option to position in one shape entities from different types. right now as demand we want to enable this option. its involved with the change of logic of existing implementation. for now the develop continues with the current option. waiting for clarity with this issue.
1.0
Position in one shape entities from different types - in the current staged app there is no option to position in one shape entities from different types. right now as demand we want to enable this option. its involved with the change of logic of existing implementation. for now the develop continues with the current option. waiting for clarity with this issue.
process
position in one shape entities from different types in the current staged app there is no option to position in one shape entities from different types right now as demand we want to enable this option its involved with the change of logic of existing implementation for now the develop continues with the current option waiting for clarity with this issue
1
236,478
19,547,835,280
IssuesEvent
2022-01-02 07:16:50
Misaka-0x447f/iis.cloudy-misaka
https://api.github.com/repos/Misaka-0x447f/iis.cloudy-misaka
opened
Set Up This Blog - Internet Information Service @ misaka.org
Gitalk https://iis.misaka.org/blog/test-post2/
https://iis.misaka.org/blog/test-post2/ This is a guide for everything to set up in the repository to make this place your own! Regardless of if you're new to Gridsome or just looking for the checklist of things to update when you clone this project, this post is for you!
1.0
Set Up This Blog - Internet Information Service @ misaka.org - https://iis.misaka.org/blog/test-post2/ This is a guide for everything to set up in the repository to make this place your own! Regardless of if you're new to Gridsome or just looking for the checklist of things to update when you clone this project, this post is for you!
non_process
set up this blog internet information service misaka org this is a guide for everything to set up in the repository to make this place your own regardless of if you re new to gridsome or just looking for the checklist of things to update when you clone this project this post is for you
0
100,305
11,185,035,697
IssuesEvent
2019-12-31 21:52:35
packethost/docs
https://api.github.com/repos/packethost/docs
closed
local-and-global-bgp.md
product-documentation
[local-and-global-bgp.md](https://github.com/packethost/docs/blob/master/products/04-network/03-advanced/02-local-and-global-bgp.md) This file is too big for its current purpose. Please work on to take out some part of it to our /guides section and link it here. For more info - @jacobsmith928
1.0
local-and-global-bgp.md - [local-and-global-bgp.md](https://github.com/packethost/docs/blob/master/products/04-network/03-advanced/02-local-and-global-bgp.md) This file is too big for its current purpose. Please work on to take out some part of it to our /guides section and link it here. For more info - @jacobsmith928
non_process
local and global bgp md this file is too big for its current purpose please work on to take out some part of it to our guides section and link it here for more info
0
274,589
23,851,930,901
IssuesEvent
2022-09-06 18:48:26
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
opened
[chip-test] chip_sw_pwrmgr_main_power_glitch_reset extension
Component:ChipLevelTest
### Test point name [chip_sw_pwrmgr_main_power_glitch_reset](https://github.com/lowRISC/opentitan/blob/620c8c0c07dc0ab05448c0e1ad707d82072f58da/hw/top_earlgrey/data/chip_testplan.hjson#L1035) ### Host side component _No response_ ### OpenTitanTool infrastructure implemented _No response_ ### Contact person @matutem, @tjaychen ### Checklist Add SVAs in order to check that: - Check that the clock valids are deasserted if a power glitch is detected. - clocks are stopped if their valid is deasserted The assertions in the second bulled point above can probably go into RTL. Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [ ] Device-side (C) component developed - [ ] Bazel build rules developed - [ ] Host-side component developed - [ ] HJSON test plan updated with test name (so it shows up in the dashboard) - [ ] Test added to dvsim nightly regression (and passing at time of checking)
1.0
[chip-test] chip_sw_pwrmgr_main_power_glitch_reset extension - ### Test point name [chip_sw_pwrmgr_main_power_glitch_reset](https://github.com/lowRISC/opentitan/blob/620c8c0c07dc0ab05448c0e1ad707d82072f58da/hw/top_earlgrey/data/chip_testplan.hjson#L1035) ### Host side component _No response_ ### OpenTitanTool infrastructure implemented _No response_ ### Contact person @matutem, @tjaychen ### Checklist Add SVAs in order to check that: - Check that the clock valids are deasserted if a power glitch is detected. - clocks are stopped if their valid is deasserted The assertions in the second bulled point above can probably go into RTL. Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [ ] Device-side (C) component developed - [ ] Bazel build rules developed - [ ] Host-side component developed - [ ] HJSON test plan updated with test name (so it shows up in the dashboard) - [ ] Test added to dvsim nightly regression (and passing at time of checking)
non_process
chip sw pwrmgr main power glitch reset extension test point name host side component no response opentitantool infrastructure implemented no response contact person matutem tjaychen checklist add svas in order to check that check that the clock valids are deasserted if a power glitch is detected clocks are stopped if their valid is deasserted the assertions in the second bulled point above can probably go into rtl please fill out this checklist as items are completed link to prs and issues as appropriate check if existing test covers most or all of this testpoint if so either extend said test to cover all points or skip the next checkboxes device side c component developed bazel build rules developed host side component developed hjson test plan updated with test name so it shows up in the dashboard test added to dvsim nightly regression and passing at time of checking
0
11,116
13,957,682,306
IssuesEvent
2020-10-24 08:07:40
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
PL: A kind request to initiate the harvesting process
Geoportal Harvesting process PL - Poland
Hi Helpdesk Team, can we kindly ask you to push harvesting process for PL catalogue? Regards, Piotr Perz
1.0
PL: A kind request to initiate the harvesting process - Hi Helpdesk Team, can we kindly ask you to push harvesting process for PL catalogue? Regards, Piotr Perz
process
pl a kind request to initiate the harvesting process hi helpdesk team can we kindly ask you to push harvesting process for pl catalogue regards piotr perz
1
21,645
30,079,567,391
IssuesEvent
2023-06-29 01:12:37
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
@pandacss/dev 0.5.0 has 1 guarddog issues
npm-silent-process-execution
```{"npm-silent-process-execution":[{"code":" (0, import_node_child_process.spawn)(import_node_process7.default.execPath, [import_node_path3.default.join(__dirname2, \"check.js\"), JSON.stringify(this.#options)], {\n detached: true,\n stdio: \"ignore\"\n }).unref();","location":"package/dist/cli-main.js:17813","message":"This package is silently executing another executable"}]}```
1.0
@pandacss/dev 0.5.0 has 1 guarddog issues - ```{"npm-silent-process-execution":[{"code":" (0, import_node_child_process.spawn)(import_node_process7.default.execPath, [import_node_path3.default.join(__dirname2, \"check.js\"), JSON.stringify(this.#options)], {\n detached: true,\n stdio: \"ignore\"\n }).unref();","location":"package/dist/cli-main.js:17813","message":"This package is silently executing another executable"}]}```
process
pandacss dev has guarddog issues npm silent process execution n detached true n stdio ignore n unref location package dist cli main js message this package is silently executing another executable
1
234,290
19,123,540,987
IssuesEvent
2021-12-01 03:00:41
libreswan/libreswan
https://api.github.com/repos/libreswan/libreswan
closed
f35: addconn needs more seccomp system calls
regression testing
``` +type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX +type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX ``` ``` +#0 __GI___fstatat64 (fd=7, file=0x7f18d8b5fff5 "", buf=0x7ffc76ee2110, flag=4096) at ../sysdeps/unix/sysv/linux/fstatat64.c:162 +#1 0x00007f18d8a1cb43 in __GI__IO_file_doallocate (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947 +#2 0x00007f18d8a2a890 in __GI__IO_doallocbuf (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947 +#3 __GI__IO_doallocbuf (fp=fp@entry=0x7f18d7d66e20) at genops.c:342 +#4 0x00007f18d8a299bc in _IO_new_file_underflow (fp=0x7f18d7d66e20) at fileops.c:486 +#5 0x00007f18d8a2a946 in __GI__IO_default_uflow (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947 +#6 0x00007f18d8a1e27c in __GI__IO_getline_info (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1, eof=eof@entry=0x0) at iogetline.c:60 +#7 0x00007f18d8a1e37c in __GI__IO_getline (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1) at iogetline.c:34 +#8 0x00007f18d8a1d360 in _IO_fgets (buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1024, fp=fp@entry=0x7f18d7d66e20) at iofgets.c:53 +#9 0x00007f18d8c5fb73 in fgets (__stream=0x7f18d7d66e20, __n=1024, __s=0x7ffc76ee2310 "\317\036") at PATH/include/bits/stdio2.h:272 +#10 ub_ctx_hosts (ctx=0x7f18d8066e70, fname=fname@entry=0x55ebeee4fbbe "/etc/hosts") at libunbound/libunbound.c:1215 +#11 0x000055ebeee3e02a in unbound_ctx_config (do_dnssec=<optimized out>, rootfile=0x7f18d8058fe0 "/var/lib/unbound/root.key", trusted=0x0, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:63 +#12 0x000055ebeee3e6bb in unbound_sync_init (do_dnssec=<optimized out>, rootfile=<optimized out>, trusted=<optimized out>, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:226 +#13 0x000055ebeee21b27 in main (argc=2, argv=0x7ffc76ee3258) at /source/programs/addconn/addconn.c:416 ```
1.0
f35: addconn needs more seccomp system calls - ``` +type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX +type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX ``` ``` +#0 __GI___fstatat64 (fd=7, file=0x7f18d8b5fff5 "", buf=0x7ffc76ee2110, flag=4096) at ../sysdeps/unix/sysv/linux/fstatat64.c:162 +#1 0x00007f18d8a1cb43 in __GI__IO_file_doallocate (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947 +#2 0x00007f18d8a2a890 in __GI__IO_doallocbuf (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947 +#3 __GI__IO_doallocbuf (fp=fp@entry=0x7f18d7d66e20) at genops.c:342 +#4 0x00007f18d8a299bc in _IO_new_file_underflow (fp=0x7f18d7d66e20) at fileops.c:486 +#5 0x00007f18d8a2a946 in __GI__IO_default_uflow (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947 +#6 0x00007f18d8a1e27c in __GI__IO_getline_info (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1, eof=eof@entry=0x0) at iogetline.c:60 +#7 0x00007f18d8a1e37c in __GI__IO_getline (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1) at iogetline.c:34 +#8 0x00007f18d8a1d360 in _IO_fgets (buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1024, fp=fp@entry=0x7f18d7d66e20) at iofgets.c:53 +#9 0x00007f18d8c5fb73 in fgets (__stream=0x7f18d7d66e20, __n=1024, __s=0x7ffc76ee2310 "\317\036") at PATH/include/bits/stdio2.h:272 +#10 ub_ctx_hosts (ctx=0x7f18d8066e70, fname=fname@entry=0x55ebeee4fbbe "/etc/hosts") at libunbound/libunbound.c:1215 +#11 0x000055ebeee3e02a in unbound_ctx_config (do_dnssec=<optimized out>, rootfile=0x7f18d8058fe0 "/var/lib/unbound/root.key", trusted=0x0, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:63 +#12 0x000055ebeee3e6bb in unbound_sync_init (do_dnssec=<optimized out>, rootfile=<optimized out>, trusted=<optimized out>, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:226 +#13 0x000055ebeee21b27 in main (argc=2, argv=0x7ffc76ee3258) at /source/programs/addconn/addconn.c:416 ```
non_process
addconn needs more seccomp system calls type seccomp msg audit xxx auid auid uid gid ses ses subj unconfined u unconfined r unconfined t pid pid comm addconn exe path libexec ipsec addconn sig arch syscall compat ip xxx type seccomp msg audit xxx auid auid uid gid ses ses subj unconfined u unconfined r unconfined t pid pid comm addconn exe path libexec ipsec addconn sig arch syscall compat ip xxx gi fd file buf flag at sysdeps unix sysv linux c in gi io file doallocate fp at path src debug glibc libio libiop h in gi io doallocbuf fp at path src debug glibc libio libiop h gi io doallocbuf fp fp entry at genops c in io new file underflow fp at fileops c in gi io default uflow fp at path src debug glibc libio libiop h in gi io getline info fp fp entry buf buf entry n n entry delim delim entry extract delim extract delim entry eof eof entry at iogetline c in gi io getline fp fp entry buf buf entry n n entry delim delim entry extract delim extract delim entry at iogetline c in io fgets buf buf entry n n entry fp fp entry at iofgets c in fgets stream n s at path include bits h ub ctx hosts ctx fname fname entry etc hosts at libunbound libunbound c in unbound ctx config do dnssec rootfile var lib unbound root key trusted logger logger entry at source lib libswan unbound c in unbound sync init do dnssec rootfile trusted logger logger entry at source lib libswan unbound c in main argc argv at source programs addconn addconn c
0
437,921
12,604,653,678
IssuesEvent
2020-06-11 15:17:07
eobanb/indianapublicmedia-web
https://api.github.com/repos/eobanb/indianapublicmedia-web
closed
WYSIWYG editor - custom formatting
enhancement medium priority
(note: as described [in Hannon Hill's documentation](https://www.hannonhill.com/cascadecms/latest/content-authoring/wysiwyg-editor-configurations/index.html)) This warrants testing / developing a proof of concept for several possible uses: - /ethergame/ hints that can be hidden/shown individually - /news/ 'view older updates' button that can be applied to a post without HTML knowledge - inline Google DFP ads (I.e. for /indiandroid/ and /theinbox/) without copying/pasting HTML blocks
1.0
WYSIWYG editor - custom formatting - (note: as described [in Hannon Hill's documentation](https://www.hannonhill.com/cascadecms/latest/content-authoring/wysiwyg-editor-configurations/index.html)) This warrants testing / developing a proof of concept for several possible uses: - /ethergame/ hints that can be hidden/shown individually - /news/ 'view older updates' button that can be applied to a post without HTML knowledge - inline Google DFP ads (I.e. for /indiandroid/ and /theinbox/) without copying/pasting HTML blocks
non_process
wysiwyg editor custom formatting note as described this warrants testing developing a proof of concept for several possible uses ethergame hints that can be hidden shown individually news view older updates button that can be applied to a post without html knowledge inline google dfp ads i e for indiandroid and theinbox without copying pasting html blocks
0
15,099
18,836,033,166
IssuesEvent
2021-11-11 01:04:19
GoogleCloudPlatform/cloud-ops-sandbox
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
opened
Ensure dependency bot is tracking terraform dependencies
type: process priority: p3
[We recently had an issue where a terraform dependency updated and broke our released code](https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/pull/864). I believe this repo should have a bot to catch these issues and auto-update dependencies (dependabot, rennovatebot, or both), but we should confirm this
1.0
Ensure dependency bot is tracking terraform dependencies - [We recently had an issue where a terraform dependency updated and broke our released code](https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/pull/864). I believe this repo should have a bot to catch these issues and auto-update dependencies (dependabot, rennovatebot, or both), but we should confirm this
process
ensure dependency bot is tracking terraform dependencies i believe this repo should have a bot to catch these issues and auto update dependencies dependabot rennovatebot or both but we should confirm this
1
11,309
14,112,001,026
IssuesEvent
2020-11-07 02:44:15
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
opened
Learning Music Helps You Read: Using Transfer to Study Linguistic Structure in Language Models
2020 Natural Language Processing Programming Languages Transfer Learning _read_later
* https://arxiv.org/abs/2004.14601 * 2020 本研究では、ニューラル言語モデルにおける文法構造の符号化を解析するための手法として、伝達学習を提案する。 本研究では、LSTMを非言語データ上で訓練し、自然言語上での性能を評価することで、どのようなデータがLSTMが自然言語で使用できる一般化可能な構造特徴を誘発するかを評価する。 その結果、非言語データに潜在的な構造を持つデータ(MIDI音楽やJavaコード)を学習させると、表面形式や語彙に重なりがないにもかかわらず、自然言語でのテスト性能が向上することがわかった。 この改善につながるモデルがどのような抽象構造をコード化しているのかを特定するために、我々は2つの人工括弧言語を用いて同様の実験を行った: 1つは階層的な再帰構造を持つ言語、もう1つはペアトークンを持つが再帰を持たない制御言語である。 驚くべきことに、これらの人工言語のいずれかでモデルを訓練すると、自然言語でテストしたときに同じような実質的な利益が得られる。 さらに、語彙の重複を制御した自然言語間の移動に関する実験では、テスト言語でのゼロショット性能は、訓練言語との類型論的な統語的類似性と高い相関があることが示された。 この結果は、ニューラルモデルが抽象的な統語構造を表現する方法や、自然言語の習得を可能にする構造的帰納的バイアスの種類についての洞察を提供する。
1.0
Learning Music Helps You Read: Using Transfer to Study Linguistic Structure in Language Models - * https://arxiv.org/abs/2004.14601 * 2020 本研究では、ニューラル言語モデルにおける文法構造の符号化を解析するための手法として、伝達学習を提案する。 本研究では、LSTMを非言語データ上で訓練し、自然言語上での性能を評価することで、どのようなデータがLSTMが自然言語で使用できる一般化可能な構造特徴を誘発するかを評価する。 その結果、非言語データに潜在的な構造を持つデータ(MIDI音楽やJavaコード)を学習させると、表面形式や語彙に重なりがないにもかかわらず、自然言語でのテスト性能が向上することがわかった。 この改善につながるモデルがどのような抽象構造をコード化しているのかを特定するために、我々は2つの人工括弧言語を用いて同様の実験を行った: 1つは階層的な再帰構造を持つ言語、もう1つはペアトークンを持つが再帰を持たない制御言語である。 驚くべきことに、これらの人工言語のいずれかでモデルを訓練すると、自然言語でテストしたときに同じような実質的な利益が得られる。 さらに、語彙の重複を制御した自然言語間の移動に関する実験では、テスト言語でのゼロショット性能は、訓練言語との類型論的な統語的類似性と高い相関があることが示された。 この結果は、ニューラルモデルが抽象的な統語構造を表現する方法や、自然言語の習得を可能にする構造的帰納的バイアスの種類についての洞察を提供する。
process
learning music helps you read using transfer to study linguistic structure in language models 本研究では、ニューラル言語モデルにおける文法構造の符号化を解析するための手法として、伝達学習を提案する。 本研究では、lstmを非言語データ上で訓練し、自然言語上での性能を評価することで、どのようなデータがlstmが自然言語で使用できる一般化可能な構造特徴を誘発するかを評価する。 その結果、非言語データに潜在的な構造を持つデータ midi音楽やjavaコード を学習させると、表面形式や語彙に重なりがないにもかかわらず、自然言語でのテスト性能が向上することがわかった。 この改善につながるモデルがどのような抽象構造をコード化しているのかを特定するために、 : 、 。 驚くべきことに、これらの人工言語のいずれかでモデルを訓練すると、自然言語でテストしたときに同じような実質的な利益が得られる。 さらに、語彙の重複を制御した自然言語間の移動に関する実験では、テスト言語でのゼロショット性能は、訓練言語との類型論的な統語的類似性と高い相関があることが示された。 この結果は、ニューラルモデルが抽象的な統語構造を表現する方法や、自然言語の習得を可能にする構造的帰納的バイアスの種類についての洞察を提供する。
1
274,622
23,853,210,457
IssuesEvent
2022-09-06 20:04:23
mountaincharlie/project-five-midi-dragon
https://api.github.com/repos/mountaincharlie/project-five-midi-dragon
closed
Testing
Testing
- [x] check the landing page renders correctly - [x] check that all styling has been properly applied - [x] create test_urls.py to test home app url - [x] create test_views.py to test home app view - [x] document tests in README
1.0
Testing - - [x] check the landing page renders correctly - [x] check that all styling has been properly applied - [x] create test_urls.py to test home app url - [x] create test_views.py to test home app view - [x] document tests in README
non_process
testing check the landing page renders correctly check that all styling has been properly applied create test urls py to test home app url create test views py to test home app view document tests in readme
0
138,383
18,785,304,236
IssuesEvent
2021-11-08 11:28:18
sultanabubaker/gradle-simple-project
https://api.github.com/repos/sultanabubaker/gradle-simple-project
reopened
CVE-2020-10673 (High) detected in jackson-databind-2.8.11.6.jar
security vulnerability
## CVE-2020-10673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: gradle-simple-project/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.6/35753201d0cdb1dbe998ab289bca1180b68d4368/jackson-databind-2.8.11.6.jar</p> <p> Dependency Hierarchy: - sami-pr-nexus-2.0.0.jar (Root Library) - core-5.0.0.jar - crypto-5.0.0.jar - :x: **jackson-databind-2.8.11.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/gradle-simple-project/commit/591721fb6f22aef1255695db7f0486533bff325a">591721fb6f22aef1255695db7f0486533bff325a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.6","packageFilePaths":["/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.jitpack:sami-pr-nexus:2.0.0;org.web3j:core:5.0.0;org.web3j:crypto:5.0.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-10673 (High) detected in jackson-databind-2.8.11.6.jar - ## CVE-2020-10673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: gradle-simple-project/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.6/35753201d0cdb1dbe998ab289bca1180b68d4368/jackson-databind-2.8.11.6.jar</p> <p> Dependency Hierarchy: - sami-pr-nexus-2.0.0.jar (Root Library) - core-5.0.0.jar - crypto-5.0.0.jar - :x: **jackson-databind-2.8.11.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/gradle-simple-project/commit/591721fb6f22aef1255695db7f0486533bff325a">591721fb6f22aef1255695db7f0486533bff325a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.6","packageFilePaths":["/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.jitpack:sami-pr-nexus:2.0.0;org.web3j:core:5.0.0;org.web3j:crypto:5.0.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file gradle simple project build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy sami pr nexus jar root library core jar crypto jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io jitpack sami pr nexus org core org crypto com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus vulnerabilityurl
0
60,995
14,939,098,501
IssuesEvent
2021-01-25 16:32:27
EIDSS/EIDSS7
https://api.github.com/repos/EIDSS/EIDSS7
closed
LAB17 Approve samples to be destroyed
Build 98.0 Minor bug
**Summary** Step 9. Verification step: in the same section enter the Lab Sample ID in the search field and select search.The approved samples records display in the Samples tab. The approved samples records display with <Sample Status> Deleted. **To Reproduce** Steps to reproduce the behavior: 1. Log in as lisasimpson 2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search. 3. The approved samples records display in the Samples tab. 4. The approved samples records display with <Sample Status> Deleted. **Expected behavior** 2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search. 3. The approved samples records display in the Samples tab. 4. The approved samples records display with <Sample Status> Destroyed. **Screenshots** ![image](https://user-images.githubusercontent.com/52667706/66413365-39428280-e9c5-11e9-941e-92e67544bc49.png) **Additional details:** - Build: - Script title (enter ad hoc if not script-based): **Issue severity (Optional)** Severity (critical, major, minor, low): **Additional context** Add any other context about the problem here.
1.0
LAB17 Approve samples to be destroyed - **Summary** Step 9. Verification step: in the same section enter the Lab Sample ID in the search field and select search.The approved samples records display in the Samples tab. The approved samples records display with <Sample Status> Deleted. **To Reproduce** Steps to reproduce the behavior: 1. Log in as lisasimpson 2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search. 3. The approved samples records display in the Samples tab. 4. The approved samples records display with <Sample Status> Deleted. **Expected behavior** 2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search. 3. The approved samples records display in the Samples tab. 4. The approved samples records display with <Sample Status> Destroyed. **Screenshots** ![image](https://user-images.githubusercontent.com/52667706/66413365-39428280-e9c5-11e9-941e-92e67544bc49.png) **Additional details:** - Build: - Script title (enter ad hoc if not script-based): **Issue severity (Optional)** Severity (critical, major, minor, low): **Additional context** Add any other context about the problem here.
non_process
approve samples to be destroyed summary step verification step in the same section enter the lab sample id in the search field and select search the approved samples records display in the samples tab the approved samples records display with deleted to reproduce steps to reproduce the behavior log in as lisasimpson go to step verification step in the same section enter the lab sample id in the search field and select search the approved samples records display in the samples tab the approved samples records display with deleted expected behavior go to step verification step in the same section enter the lab sample id in the search field and select search the approved samples records display in the samples tab the approved samples records display with destroyed screenshots additional details build script title enter ad hoc if not script based issue severity optional severity critical major minor low additional context add any other context about the problem here
0
4,875
7,753,598,812
IssuesEvent
2018-05-31 01:38:07
kubernetes-sigs/kubebuilder
https://api.github.com/repos/kubernetes-sigs/kubebuilder
closed
Create a SECURITY_CONTACTS file.
process
As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS file. The template for the file can be found in the kubernetes-template repository[2]. A description for the file is in the steering-committee docs[3], you might need to search that page for "Security Contacts". Please feel free to ping me on the PR when you make it, otherwise I will see when you close this issue. :) Thanks so much, let me know if you have any questions. (This issue was generated from a tool, apologies for any weirdness.) [1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE [2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS [3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
1.0
Create a SECURITY_CONTACTS file. - As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS file. The template for the file can be found in the kubernetes-template repository[2]. A description for the file is in the steering-committee docs[3], you might need to search that page for "Security Contacts". Please feel free to ping me on the PR when you make it, otherwise I will see when you close this issue. :) Thanks so much, let me know if you have any questions. (This issue was generated from a tool, apologies for any weirdness.) [1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE [2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS [3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
process
create a security contacts file as per the email sent to kubernetes dev please create a security contacts file the template for the file can be found in the kubernetes template repository a description for the file is in the steering committee docs you might need to search that page for security contacts please feel free to ping me on the pr when you make it otherwise i will see when you close this issue thanks so much let me know if you have any questions this issue was generated from a tool apologies for any weirdness
1
106,148
4,263,799,489
IssuesEvent
2016-07-12 02:45:34
ampproject/docs
https://api.github.com/repos/ampproject/docs
closed
Invalid word breaking
Priority: High
Word breaking is invalid, unacceptable. The culprit must be hiding somewhere [in these lines](https://github.com/ampproject/docs/blob/5ce07f344ad482d23e34032e052d3849861cd6ba/_sass/content.scss#L42). ![what is amp 2016-03-02 20-37-04](https://cloud.githubusercontent.com/assets/684364/13472870/0ddf886a-e0b7-11e5-937a-59dbc7484265.png)
1.0
Invalid word breaking - Word breaking is invalid, unacceptable. The culprit must be hiding somewhere [in these lines](https://github.com/ampproject/docs/blob/5ce07f344ad482d23e34032e052d3849861cd6ba/_sass/content.scss#L42). ![what is amp 2016-03-02 20-37-04](https://cloud.githubusercontent.com/assets/684364/13472870/0ddf886a-e0b7-11e5-937a-59dbc7484265.png)
non_process
invalid word breaking word breaking is invalid unacceptable the culprit must be hiding somewhere
0
1,827
4,613,585,929
IssuesEvent
2016-09-25 03:25:34
EBrown8534/StackExchangeStatisticsExplorer
https://api.github.com/repos/EBrown8534/StackExchangeStatisticsExplorer
closed
Going to `/Sites/Compare/` with no `Sites` Query String parameter throws error
bug in process
Self explanatory, if you go to the `/Sites/Compare/` endpoint without a valid `Sites` query string parameter it should redirect to `/Sites/Select/` just like `/Sites/Detail/` redirects to `/Sites/Default/` without a valid `Site` query string parameter.
1.0
Going to `/Sites/Compare/` with no `Sites` Query String parameter throws error - Self explanatory, if you go to the `/Sites/Compare/` endpoint without a valid `Sites` query string parameter it should redirect to `/Sites/Select/` just like `/Sites/Detail/` redirects to `/Sites/Default/` without a valid `Site` query string parameter.
process
going to sites compare with no sites query string parameter throws error self explanatory if you go to the sites compare endpoint without a valid sites query string parameter it should redirect to sites select just like sites detail redirects to sites default without a valid site query string parameter
1
4,856
7,745,641,995
IssuesEvent
2018-05-29 18:59:54
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
The invoke does not work.
cxp in-process storage/svc triaged
Invoke-StorageSyncFileRecall : The term 'Invoke-StorageSyncFileRecall' is not recognized as the name of a cmdlet. This is despite having ran the Import-Module. Also, on the Import-Module command for StorageSync.Management.ServerCmdlets.dll complains about unapproved verbs. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 4814031f-3cca-10a6-8587-cf1a045d2481 * Version Independent ID: b96d1378-9a11-c00e-1c84-a69d4ff1cff3 * Content: [Manage registered servers with Azure File Sync (preview)](https://docs.microsoft.com/en-us/azure/storage/files/storage-sync-files-server-registration#feedback) * Content Source: [articles/storage/files/storage-sync-files-server-registration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/storage/files/storage-sync-files-server-registration.md) * Service: **storage** * Product: **unspecified** * GitHub Login: @wmgries * Microsoft Alias: **wgries**
1.0
The invoke does not work. - Invoke-StorageSyncFileRecall : The term 'Invoke-StorageSyncFileRecall' is not recognized as the name of a cmdlet. This is despite having ran the Import-Module. Also, on the Import-Module command for StorageSync.Management.ServerCmdlets.dll complains about unapproved verbs. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 4814031f-3cca-10a6-8587-cf1a045d2481 * Version Independent ID: b96d1378-9a11-c00e-1c84-a69d4ff1cff3 * Content: [Manage registered servers with Azure File Sync (preview)](https://docs.microsoft.com/en-us/azure/storage/files/storage-sync-files-server-registration#feedback) * Content Source: [articles/storage/files/storage-sync-files-server-registration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/storage/files/storage-sync-files-server-registration.md) * Service: **storage** * Product: **unspecified** * GitHub Login: @wmgries * Microsoft Alias: **wgries**
process
the invoke does not work invoke storagesyncfilerecall the term invoke storagesyncfilerecall is not recognized as the name of a cmdlet this is despite having ran the import module also on the import module command for storagesync management servercmdlets dll complains about unapproved verbs document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service storage product unspecified github login wmgries microsoft alias wgries
1
21,590
29,983,918,107
IssuesEvent
2023-06-25 02:00:07
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Fri, 23 Jun 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Exploring the Role of Audio in Video Captioning - **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2306.12559 - **Pdf link:** https://arxiv.org/pdf/2306.12559 - **Abstract** Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Data-Free Backbone Fine-Tuning for Pruned Neural Networks - **Authors:** Adrian Holzbock, Achyut Hegde, Klaus Dietmayer, Vasileios Belagiannis - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.12881 - **Pdf link:** https://arxiv.org/pdf/2306.12881 - **Abstract** Model compression techniques reduce the computational load and memory consumption of deep neural networks. After the compression operation, e.g. parameter pruning, the model is normally fine-tuned on the original training dataset to recover from the performance drop caused by compression. However, the training data is not always available due to privacy issues or other factors. In this work, we present a data-free fine-tuning approach for pruning the backbone of deep neural networks. In particular, the pruned network backbone is trained with synthetically generated images, and our proposed intermediate supervision to mimic the unpruned backbone's output feature map. Afterwards, the pruned backbone can be combined with the original network head to make predictions. We generate synthetic images by back-propagating gradients to noise images while relying on L1-pruning for the backbone pruning. In our experiments, we show that our approach is task-independent due to pruning only the backbone. By evaluating our approach on 2D human pose estimation, object detection, and image classification, we demonstrate promising performance compared to the unpruned model. Our code is available at https://github.com/holzbock/dfbf. ## Keyword: RAW ### Exploring the Role of Audio in Video Captioning - **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2306.12559 - **Pdf link:** https://arxiv.org/pdf/2306.12559 - **Abstract** Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input. ### Neural Spectro-polarimetric Fields - **Authors:** Youngchan Kim, Wonjoon Jin, Sunghyun Cho, Seung-Hwan Baek - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12562 - **Pdf link:** https://arxiv.org/pdf/2306.12562 - **Abstract** Modeling the spatial radiance distribution of light rays in a scene has been extensively explored for applications, including view synthesis. Spectrum and polarization, the wave properties of light, are often neglected due to their integration into three RGB spectral bands and their non-perceptibility to human vision. Despite this, these properties encompass substantial material and geometric information about a scene. In this work, we propose to model spectro-polarimetric fields, the spatial Stokes-vector distribution of any light ray at an arbitrary wavelength. We present Neural Spectro-polarimetric Fields (NeSpoF), a neural representation that models the physically-valid Stokes vector at given continuous variables of position, direction, and wavelength. NeSpoF manages inherently noisy raw measurements, showcases memory efficiency, and preserves physically vital signals, factors that are crucial for representing the high-dimensional signal of a spectro-polarimetric field. To validate NeSpoF, we introduce the first multi-view hyperspectral-polarimetric image dataset, comprised of both synthetic and real-world scenes. These were captured using our compact hyperspectral-polarimetric imaging system, which has been calibrated for robustness against system imperfections. We demonstrate the capabilities of NeSpoF on diverse scenes. ### DreamEdit: Subject-driven Image Editing - **Authors:** Tianle Li, Max Ku, Cong Wei, Wenhu Chen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.12624 - **Pdf link:** https://arxiv.org/pdf/2306.12624 - **Abstract** Subject-driven image generation aims at generating images containing customized subjects, which has recently drawn enormous attention from the research community. However, the previous works cannot precisely control the background and position of the target subject. In this work, we aspire to fill the void and propose two novel subject-driven sub-tasks, i.e., Subject Replacement and Subject Addition. The new tasks are challenging in multiple aspects: replacing a subject with a customized one can change its shape, texture, and color, while adding a target subject to a designated position in a provided scene necessitates a context-aware posture. To conquer these two novel tasks, we first manually curate a new dataset DreamEditBench containing 22 different types of subjects, and 440 source images with different difficulty levels. We plan to host DreamEditBench as a platform and hire trained evaluators for standard human evaluation. We also devise an innovative method DreamEditor to resolve these tasks by performing iterative generation, which enables a smooth adaptation to the customized subject. In this project, we conduct automatic and human evaluations to understand the performance of DreamEditor and baselines on DreamEditBench. For Subject Replacement, we found that the existing models are sensitive to the shape and color of the original subject. The model failure rate will dramatically increase when the source and target subjects are highly different. For Subject Addition, we found that the existing models cannot easily blend the customized subjects into the background smoothly, leading to noticeable artifacts in the generated image. We hope DreamEditBench can become a standard platform to enable future investigations toward building more controllable subject-driven image editing. Our project homepage is https://dreameditbenchteam.github.io/. ## Keyword: raw image There is no result
2.0
New submissions for Fri, 23 Jun 23 - ## Keyword: events ### Exploring the Role of Audio in Video Captioning - **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2306.12559 - **Pdf link:** https://arxiv.org/pdf/2306.12559 - **Abstract** Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Data-Free Backbone Fine-Tuning for Pruned Neural Networks - **Authors:** Adrian Holzbock, Achyut Hegde, Klaus Dietmayer, Vasileios Belagiannis - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.12881 - **Pdf link:** https://arxiv.org/pdf/2306.12881 - **Abstract** Model compression techniques reduce the computational load and memory consumption of deep neural networks. After the compression operation, e.g. parameter pruning, the model is normally fine-tuned on the original training dataset to recover from the performance drop caused by compression. However, the training data is not always available due to privacy issues or other factors. In this work, we present a data-free fine-tuning approach for pruning the backbone of deep neural networks. In particular, the pruned network backbone is trained with synthetically generated images, and our proposed intermediate supervision to mimic the unpruned backbone's output feature map. Afterwards, the pruned backbone can be combined with the original network head to make predictions. We generate synthetic images by back-propagating gradients to noise images while relying on L1-pruning for the backbone pruning. In our experiments, we show that our approach is task-independent due to pruning only the backbone. By evaluating our approach on 2D human pose estimation, object detection, and image classification, we demonstrate promising performance compared to the unpruned model. Our code is available at https://github.com/holzbock/dfbf. ## Keyword: RAW ### Exploring the Role of Audio in Video Captioning - **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2306.12559 - **Pdf link:** https://arxiv.org/pdf/2306.12559 - **Abstract** Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input. ### Neural Spectro-polarimetric Fields - **Authors:** Youngchan Kim, Wonjoon Jin, Sunghyun Cho, Seung-Hwan Baek - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12562 - **Pdf link:** https://arxiv.org/pdf/2306.12562 - **Abstract** Modeling the spatial radiance distribution of light rays in a scene has been extensively explored for applications, including view synthesis. Spectrum and polarization, the wave properties of light, are often neglected due to their integration into three RGB spectral bands and their non-perceptibility to human vision. Despite this, these properties encompass substantial material and geometric information about a scene. In this work, we propose to model spectro-polarimetric fields, the spatial Stokes-vector distribution of any light ray at an arbitrary wavelength. We present Neural Spectro-polarimetric Fields (NeSpoF), a neural representation that models the physically-valid Stokes vector at given continuous variables of position, direction, and wavelength. NeSpoF manages inherently noisy raw measurements, showcases memory efficiency, and preserves physically vital signals, factors that are crucial for representing the high-dimensional signal of a spectro-polarimetric field. To validate NeSpoF, we introduce the first multi-view hyperspectral-polarimetric image dataset, comprised of both synthetic and real-world scenes. These were captured using our compact hyperspectral-polarimetric imaging system, which has been calibrated for robustness against system imperfections. We demonstrate the capabilities of NeSpoF on diverse scenes. ### DreamEdit: Subject-driven Image Editing - **Authors:** Tianle Li, Max Ku, Cong Wei, Wenhu Chen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.12624 - **Pdf link:** https://arxiv.org/pdf/2306.12624 - **Abstract** Subject-driven image generation aims at generating images containing customized subjects, which has recently drawn enormous attention from the research community. However, the previous works cannot precisely control the background and position of the target subject. In this work, we aspire to fill the void and propose two novel subject-driven sub-tasks, i.e., Subject Replacement and Subject Addition. The new tasks are challenging in multiple aspects: replacing a subject with a customized one can change its shape, texture, and color, while adding a target subject to a designated position in a provided scene necessitates a context-aware posture. To conquer these two novel tasks, we first manually curate a new dataset DreamEditBench containing 22 different types of subjects, and 440 source images with different difficulty levels. We plan to host DreamEditBench as a platform and hire trained evaluators for standard human evaluation. We also devise an innovative method DreamEditor to resolve these tasks by performing iterative generation, which enables a smooth adaptation to the customized subject. In this project, we conduct automatic and human evaluations to understand the performance of DreamEditor and baselines on DreamEditBench. For Subject Replacement, we found that the existing models are sensitive to the shape and color of the original subject. The model failure rate will dramatically increase when the source and target subjects are highly different. For Subject Addition, we found that the existing models cannot easily blend the customized subjects into the background smoothly, leading to noticeable artifacts in the generated image. We hope DreamEditBench can become a standard platform to enable future investigations toward building more controllable subject-driven image editing. Our project homepage is https://dreameditbenchteam.github.io/. ## Keyword: raw image There is no result
process
new submissions for fri jun keyword events exploring the role of audio in video captioning authors yuhan shen linjie yang longyin wen haichao yu ehsan elhamifar heng wang subjects computer vision and pattern recognition cs cv sound cs sd audio and speech processing eess as arxiv link pdf link abstract recent focus in video captioning has been on designing architectures that can consume both video and text modalities and using large scale video datasets with text transcripts for pre training such as though these approaches have achieved significant improvement the audio modality is often ignored in video captioning in this work we present an audio visual framework which aims to fully exploit the potential of the audio modality for captioning instead of relying on text transcripts extracted via automatic speech recognition asr we argue that learning with raw audio signals can be more beneficial as audio has additional information including acoustic events speaker identity etc our contributions are twofold first we observed that the model overspecializes to the audio modality when pre training with both video and audio modality since the ground truth i e text transcripts can be solely predicted using audio we proposed a modality balanced pre training mbp loss to mitigate this issue and significantly improve the performance on downstream tasks second we slice and dice different design choices of the cross modal module which may become an information bottleneck and generate inferior results we proposed new local global fusion mechanisms to improve information exchange across audio and video we demonstrate significant improvements by leveraging the audio modality on four datasets and even outperform the state of the art on some metrics without relying on the text modality as the input keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression data free backbone fine tuning for pruned neural networks authors adrian holzbock achyut hegde klaus dietmayer vasileios belagiannis subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract model compression techniques reduce the computational load and memory consumption of deep neural networks after the compression operation e g parameter pruning the model is normally fine tuned on the original training dataset to recover from the performance drop caused by compression however the training data is not always available due to privacy issues or other factors in this work we present a data free fine tuning approach for pruning the backbone of deep neural networks in particular the pruned network backbone is trained with synthetically generated images and our proposed intermediate supervision to mimic the unpruned backbone s output feature map afterwards the pruned backbone can be combined with the original network head to make predictions we generate synthetic images by back propagating gradients to noise images while relying on pruning for the backbone pruning in our experiments we show that our approach is task independent due to pruning only the backbone by evaluating our approach on human pose estimation object detection and image classification we demonstrate promising performance compared to the unpruned model our code is available at keyword raw exploring the role of audio in video captioning authors yuhan shen linjie yang longyin wen haichao yu ehsan elhamifar heng wang subjects computer vision and pattern recognition cs cv sound cs sd audio and speech processing eess as arxiv link pdf link abstract recent focus in video captioning has been on designing architectures that can consume both video and text modalities and using large scale video datasets with text transcripts for pre training such as though these approaches have achieved significant improvement the audio modality is often ignored in video captioning in this work we present an audio visual framework which aims to fully exploit the potential of the audio modality for captioning instead of relying on text transcripts extracted via automatic speech recognition asr we argue that learning with raw audio signals can be more beneficial as audio has additional information including acoustic events speaker identity etc our contributions are twofold first we observed that the model overspecializes to the audio modality when pre training with both video and audio modality since the ground truth i e text transcripts can be solely predicted using audio we proposed a modality balanced pre training mbp loss to mitigate this issue and significantly improve the performance on downstream tasks second we slice and dice different design choices of the cross modal module which may become an information bottleneck and generate inferior results we proposed new local global fusion mechanisms to improve information exchange across audio and video we demonstrate significant improvements by leveraging the audio modality on four datasets and even outperform the state of the art on some metrics without relying on the text modality as the input neural spectro polarimetric fields authors youngchan kim wonjoon jin sunghyun cho seung hwan baek subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract modeling the spatial radiance distribution of light rays in a scene has been extensively explored for applications including view synthesis spectrum and polarization the wave properties of light are often neglected due to their integration into three rgb spectral bands and their non perceptibility to human vision despite this these properties encompass substantial material and geometric information about a scene in this work we propose to model spectro polarimetric fields the spatial stokes vector distribution of any light ray at an arbitrary wavelength we present neural spectro polarimetric fields nespof a neural representation that models the physically valid stokes vector at given continuous variables of position direction and wavelength nespof manages inherently noisy raw measurements showcases memory efficiency and preserves physically vital signals factors that are crucial for representing the high dimensional signal of a spectro polarimetric field to validate nespof we introduce the first multi view hyperspectral polarimetric image dataset comprised of both synthetic and real world scenes these were captured using our compact hyperspectral polarimetric imaging system which has been calibrated for robustness against system imperfections we demonstrate the capabilities of nespof on diverse scenes dreamedit subject driven image editing authors tianle li max ku cong wei wenhu chen subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract subject driven image generation aims at generating images containing customized subjects which has recently drawn enormous attention from the research community however the previous works cannot precisely control the background and position of the target subject in this work we aspire to fill the void and propose two novel subject driven sub tasks i e subject replacement and subject addition the new tasks are challenging in multiple aspects replacing a subject with a customized one can change its shape texture and color while adding a target subject to a designated position in a provided scene necessitates a context aware posture to conquer these two novel tasks we first manually curate a new dataset dreameditbench containing different types of subjects and source images with different difficulty levels we plan to host dreameditbench as a platform and hire trained evaluators for standard human evaluation we also devise an innovative method dreameditor to resolve these tasks by performing iterative generation which enables a smooth adaptation to the customized subject in this project we conduct automatic and human evaluations to understand the performance of dreameditor and baselines on dreameditbench for subject replacement we found that the existing models are sensitive to the shape and color of the original subject the model failure rate will dramatically increase when the source and target subjects are highly different for subject addition we found that the existing models cannot easily blend the customized subjects into the background smoothly leading to noticeable artifacts in the generated image we hope dreameditbench can become a standard platform to enable future investigations toward building more controllable subject driven image editing our project homepage is keyword raw image there is no result
1
17,871
5,529,950,527
IssuesEvent
2017-03-21 00:24:42
bcgov/api-specs
https://api.github.com/repos/bcgov/api-specs
closed
Auto generate localities where HWY 97 doesn't have street directions
data prep scripts enhancement GEOCODER medium priority
Graham to generate a spreadsheet of in the form required by the address prep scripts. Michael to review and send back to Graham who will then integrate the locality list into the address prep scripts.
1.0
Auto generate localities where HWY 97 doesn't have street directions - Graham to generate a spreadsheet of in the form required by the address prep scripts. Michael to review and send back to Graham who will then integrate the locality list into the address prep scripts.
non_process
auto generate localities where hwy doesn t have street directions graham to generate a spreadsheet of in the form required by the address prep scripts michael to review and send back to graham who will then integrate the locality list into the address prep scripts
0
118,060
25,243,222,077
IssuesEvent
2022-11-15 09:12:52
arduino/arduino-cli
https://api.github.com/repos/arduino/arduino-cli
closed
Boost libraries with the exact match
type: enhancement topic: code
### Describe the request > When I search for SD I expect to see [SD](https://github.com/arduino-libraries/SD) to show up in the top of the list so I can install it without a lot of scrolling through the list of results. But instead it is far down in the results. Related to https://github.com/arduino/arduino-ide/issues/1106 ### Describe the current behavior `arduino-cli lib search SD` [out.txt](https://github.com/arduino/arduino-cli/files/10005220/out.txt) As you can see the SD lib is pretty down in the list ### Arduino CLI version 0.28.0 ### Operating system N/A ### Operating system version na ### Additional context _No response_ ### Issue checklist - [X] I searched for previous requests in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=) - [X] I verified the feature was still missing when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds) - [X] My request contains all necessary details
1.0
Boost libraries with the exact match - ### Describe the request > When I search for SD I expect to see [SD](https://github.com/arduino-libraries/SD) to show up in the top of the list so I can install it without a lot of scrolling through the list of results. But instead it is far down in the results. Related to https://github.com/arduino/arduino-ide/issues/1106 ### Describe the current behavior `arduino-cli lib search SD` [out.txt](https://github.com/arduino/arduino-cli/files/10005220/out.txt) As you can see the SD lib is pretty down in the list ### Arduino CLI version 0.28.0 ### Operating system N/A ### Operating system version na ### Additional context _No response_ ### Issue checklist - [X] I searched for previous requests in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=) - [X] I verified the feature was still missing when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds) - [X] My request contains all necessary details
non_process
boost libraries with the exact match describe the request when i search for sd i expect to see to show up in the top of the list so i can install it without a lot of scrolling through the list of results but instead it is far down in the results related to describe the current behavior arduino cli lib search sd as you can see the sd lib is pretty down in the list arduino cli version operating system n a operating system version na additional context no response issue checklist i searched for previous requests in i verified the feature was still missing when using the my request contains all necessary details
0
35,884
2,793,805,886
IssuesEvent
2015-05-11 13:32:23
SarineTechnologies/sarine.viewer.templates
https://api.github.com/repos/SarineTechnologies/sarine.viewer.templates
opened
Design issues
High priority Template v.1.10.0
issues to resolve: https://projects.invisionapp.com/share/MG276SMQW#/screens 1. "Hearts & Arrows" at the top should be broken to 2 lines. 2. Distance between (at the top) H&A and cut should be shortened. 3. Question mark should be blue (it may be blue, i am unable to determine if it is blue or purple). 4. In the cut screen, text should be "View the accurate geometrical measurements of the diamond to evaluate the proportions." (not like in the design).
1.0
Design issues - issues to resolve: https://projects.invisionapp.com/share/MG276SMQW#/screens 1. "Hearts & Arrows" at the top should be broken to 2 lines. 2. Distance between (at the top) H&A and cut should be shortened. 3. Question mark should be blue (it may be blue, i am unable to determine if it is blue or purple). 4. In the cut screen, text should be "View the accurate geometrical measurements of the diamond to evaluate the proportions." (not like in the design).
non_process
design issues issues to resolve hearts arrows at the top should be broken to lines distance between at the top h a and cut should be shortened question mark should be blue it may be blue i am unable to determine if it is blue or purple in the cut screen text should be view the accurate geometrical measurements of the diamond to evaluate the proportions not like in the design
0
9,776
12,794,135,735
IssuesEvent
2020-07-02 06:10:59
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Linux and Windows Service Containers
Pri2 devops-cicd-process/tech devops/prod product-question
Our team wants to run steps within a Linux container that uses a Windows service container (https://hub.docker.com/r/microsoft/azure-cosmosdb-emulator) Do you have any advice for how we can achieve this today? Is it something that will be possible in the future? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7d1c794c-d676-fc7e-758e-3464f16a1554 * Version Independent ID: 941de04a-2d34-ec82-8612-00e4070de673 * Content: [Service Containers - Azure Pipelines & TFS](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/service-containers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/service-containers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/service-containers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Linux and Windows Service Containers - Our team wants to run steps within a Linux container that uses a Windows service container (https://hub.docker.com/r/microsoft/azure-cosmosdb-emulator) Do you have any advice for how we can achieve this today? Is it something that will be possible in the future? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7d1c794c-d676-fc7e-758e-3464f16a1554 * Version Independent ID: 941de04a-2d34-ec82-8612-00e4070de673 * Content: [Service Containers - Azure Pipelines & TFS](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/service-containers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/service-containers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/service-containers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
linux and windows service containers our team wants to run steps within a linux container that uses a windows service container do you have any advice for how we can achieve this today is it something that will be possible in the future document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
5,392
8,214,905,781
IssuesEvent
2018-09-05 02:11:09
googleapis/nodejs-translate
https://api.github.com/repos/googleapis/nodejs-translate
closed
There no named export "Translate"
type: process
Here => https://github.com/googleapis/nodejs-translate/blob/master/samples/quickstart.js#L20 The quickstart is wrong, it should be: ``` const Translate = require('@google-cloud/translate'); ```
1.0
There no named export "Translate" - Here => https://github.com/googleapis/nodejs-translate/blob/master/samples/quickstart.js#L20 The quickstart is wrong, it should be: ``` const Translate = require('@google-cloud/translate'); ```
process
there no named export translate here the quickstart is wrong it should be const translate require google cloud translate
1
486,308
14,007,165,724
IssuesEvent
2020-10-28 21:07:03
AOEpeople/crawler
https://api.github.com/repos/AOEpeople/crawler
opened
[BUG] Crawler return invalid content/cache to StaticFileCache
3rd party ext Bug Priority 1 TYPO3v10 TYPO3v9 refactoring
## Bug Report This bug is still WIP in regard to describing it. But to not lose the problem I'll write down what I have for now. There is an issue on the Static File Cache GitHub on Crawler Compatibility: https://github.com/lochmueller/staticfilecache/issues/260 The crawler appears to have some issues with the Middleware Handling and the content/caching that is handled over to the StaticFileCache. This is currently resulting in an Invalid cache in Static File Cache. This is since commit: https://github.com/lochmueller/staticfilecache/commit/975eff6b95c712e34df6ff8faad8be20ef51032a omitted by a warning in regard to the crawler. As we don't want to break functionality of other extensions, and of course not lose users of the Crawler we will try to have this fixed. If you have any information that could be helpful to solve this issue, please add a comment below and lets see how we can best solve this issue.
1.0
[BUG] Crawler return invalid content/cache to StaticFileCache - ## Bug Report This bug is still WIP in regard to describing it. But to not lose the problem I'll write down what I have for now. There is an issue on the Static File Cache GitHub on Crawler Compatibility: https://github.com/lochmueller/staticfilecache/issues/260 The crawler appears to have some issues with the Middleware Handling and the content/caching that is handled over to the StaticFileCache. This is currently resulting in an Invalid cache in Static File Cache. This is since commit: https://github.com/lochmueller/staticfilecache/commit/975eff6b95c712e34df6ff8faad8be20ef51032a omitted by a warning in regard to the crawler. As we don't want to break functionality of other extensions, and of course not lose users of the Crawler we will try to have this fixed. If you have any information that could be helpful to solve this issue, please add a comment below and lets see how we can best solve this issue.
non_process
crawler return invalid content cache to staticfilecache bug report this bug is still wip in regard to describing it but to not lose the problem i ll write down what i have for now there is an issue on the static file cache github on crawler compatibility the crawler appears to have some issues with the middleware handling and the content caching that is handled over to the staticfilecache this is currently resulting in an invalid cache in static file cache this is since commit omitted by a warning in regard to the crawler as we don t want to break functionality of other extensions and of course not lose users of the crawler we will try to have this fixed if you have any information that could be helpful to solve this issue please add a comment below and lets see how we can best solve this issue
0
2,200
5,040,616,618
IssuesEvent
2016-12-19 06:41:53
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
opened
[subtitles] [eng] A bas les paradis fiscaux ! J L Mélenchon au Parlement européen
Language: English Process: [1] Writing in progress
# Video title A bas les paradis fiscaux ! J L Mélenchon au Parlement européen # URL https://www.youtube.com/watch?v=2TbjWeOxQNg&t=4s Youtube subtitle language Anglais Duration 1:08 URL subtitles https://www.youtube.com/timedtext_editor?lang=en&v=2TbjWeOxQNg&tab=captions&ref=player&action_mde_edit_form=1&ui=hd&bl=vmp
1.0
[subtitles] [eng] A bas les paradis fiscaux ! J L Mélenchon au Parlement européen - # Video title A bas les paradis fiscaux ! J L Mélenchon au Parlement européen # URL https://www.youtube.com/watch?v=2TbjWeOxQNg&t=4s Youtube subtitle language Anglais Duration 1:08 URL subtitles https://www.youtube.com/timedtext_editor?lang=en&v=2TbjWeOxQNg&tab=captions&ref=player&action_mde_edit_form=1&ui=hd&bl=vmp
process
a bas les paradis fiscaux j l mélenchon au parlement européen video title a bas les paradis fiscaux j l mélenchon au parlement européen url youtube subtitle language anglais duration url subtitles
1