Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
169
| 2,586,826,095
|
IssuesEvent
|
2015-02-17 14:49:49
|
MozillaFoundation/plan
|
https://api.github.com/repos/MozillaFoundation/plan
|
closed
|
Prototype a better etherpad for MoFos
|
p2 process
|
Add support for images, formatting and more.
### RACI
* Phase: Prototyping
* Owner: @openmatt
* Decision Maker: @openmatt
* Design Lead: not needed
* Development Lead: @jdotpz
* Quality Verifier: @davidascher
* Issues: https://github.com/MozillaFoundation/mofo-lightsaber/issues
MoFos use etherpads every day. Let's try to give them a better one. With support for things like images, formatting, better exporting to HTML and markdown, etc.
### Audience
* MoFos + the community and people they work with
### Success
* Formatting
* Images
* New ways to find / track pads
* Embedding pads. through an iframe, etc.
* Better exporting. So that etherpads can be turned into posts, pages, PDFs and tickets, etc.
* Stretch goal: make etherpad better so that people aren't as inclined to use Google Docs. Which are hard to find, require permission-setting, and not as open.
### Vision
* Updating to the latest instance and experimenting with the right plug-ins should be enough to test.
### Measurement
* Success will be measured by happier MoFos. But also: improved behavior. On how we create and share documents and drafts in progress.
* More visual thinking. Better visual style and visual language in our communication. Easier passing from etherpads into more readable / hardened artefacts and documents.
|
1.0
|
Prototype a better etherpad for MoFos - Add support for images, formatting and more.
### RACI
* Phase: Prototyping
* Owner: @openmatt
* Decision Maker: @openmatt
* Design Lead: not needed
* Development Lead: @jdotpz
* Quality Verifier: @davidascher
* Issues: https://github.com/MozillaFoundation/mofo-lightsaber/issues
MoFos use etherpads every day. Let's try to give them a better one. With support for things like images, formatting, better exporting to HTML and markdown, etc.
### Audience
* MoFos + the community and people they work with
### Success
* Formatting
* Images
* New ways to find / track pads
* Embedding pads. through an iframe, etc.
* Better exporting. So that etherpads can be turned into posts, pages, PDFs and tickets, etc.
* Stretch goal: make etherpad better so that people aren't as inclined to use Google Docs. Which are hard to find, require permission-setting, and not as open.
### Vision
* Updating to the latest instance and experimenting with the right plug-ins should be enough to test.
### Measurement
* Success will be measured by happier MoFos. But also: improved behavior. On how we create and share documents and drafts in progress.
* More visual thinking. Better visual style and visual language in our communication. Easier passing from etherpads into more readable / hardened artefacts and documents.
|
process
|
prototype a better etherpad for mofos add support for images formatting and more raci phase prototyping owner openmatt decision maker openmatt design lead not needed development lead jdotpz quality verifier davidascher issues mofos use etherpads every day let s try to give them a better one with support for things like images formatting better exporting to html and markdown etc audience mofos the community and people they work with success formatting images new ways to find track pads embedding pads through an iframe etc better exporting so that etherpads can be turned into posts pages pdfs and tickets etc stretch goal make etherpad better so that people aren t as inclined to use google docs which are hard to find require permission setting and not as open vision updating to the latest instance and experimenting with the right plug ins should be enough to test measurement success will be measured by happier mofos but also improved behavior on how we create and share documents and drafts in progress more visual thinking better visual style and visual language in our communication easier passing from etherpads into more readable hardened artefacts and documents
| 1
|
41,252
| 16,675,835,268
|
IssuesEvent
|
2021-06-07 16:03:54
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[Alerting] [o11y] Determine how to represent collected metrics
|
Feature:Actions Feature:Alerting Feature:Task Manager Team:Alerting Services
|
Relates to https://github.com/elastic/kibana/issues/98902#issuecomment-840698137
Once we determine the set of metrics that are valuable (see https://github.com/elastic/kibana/issues/100675), we want to take a small amount of time to think through how an end user might leverage these metrics. This will most likely be some sort of dashboard, but it will be helpful to know as many specifics as possible to ensure the data is shaped appropriately.
For example, if we envision users will only (at least for the short term) be interested in the top n slowest of something, perhaps we can modify data collection to only collect those particular data points (as our current [monitoring metrics](https://github.com/elastic/kibana/blob/master/x-pack/plugins/task_manager/server/MONITORING.md) collect rolling averages so it seems possible to do a little logic before shipping data to only ship top n slowest, or something along those lines)
|
1.0
|
[Alerting] [o11y] Determine how to represent collected metrics - Relates to https://github.com/elastic/kibana/issues/98902#issuecomment-840698137
Once we determine the set of metrics that are valuable (see https://github.com/elastic/kibana/issues/100675), we want to take a small amount of time to think through how an end user might leverage these metrics. This will most likely be some sort of dashboard, but it will be helpful to know as many specifics as possible to ensure the data is shaped appropriately.
For example, if we envision users will only (at least for the short term) be interested in the top n slowest of something, perhaps we can modify data collection to only collect those particular data points (as our current [monitoring metrics](https://github.com/elastic/kibana/blob/master/x-pack/plugins/task_manager/server/MONITORING.md) collect rolling averages so it seems possible to do a little logic before shipping data to only ship top n slowest, or something along those lines)
|
non_process
|
determine how to represent collected metrics relates to once we determine the set of metrics that are valuable see we want to take a small amount of time to think through how an end user might leverage these metrics this will most likely be some sort of dashboard but it will be helpful to know as many specifics as possible to ensure the data is shaped appropriately for example if we envision users will only at least for the short term be interested in the top n slowest of something perhaps we can modify data collection to only collect those particular data points as our current collect rolling averages so it seems possible to do a little logic before shipping data to only ship top n slowest or something along those lines
| 0
|
236,138
| 18,072,033,597
|
IssuesEvent
|
2021-09-21 04:45:45
|
girlscript/winter-of-contributing
|
https://api.github.com/repos/girlscript/winter-of-contributing
|
closed
|
JS: Execution Context in JavaScript
|
documentation GWOC21 JavaScript Assigned
|
<hr>
## Description 📜
I'll be explaining briefly about Execution Context in JavaScript in the documentation.
<hr>
## Domain of Contribution 📊
- [x] JavaScript
<hr>
## Directory
`winter-of-contributing/Web_Development/Javascript/Topics/Execution_Context`
<hr>
## Formats
- [x] Document
<hr>
I am a GWOC'21 contributor from Batch 1 of JavaScript. Please assign this issue to me. I assure you that this issue's title will be the same as my PR's title as well as I will be Contributing under the Guidelines and Code of Conduct.
|
1.0
|
JS: Execution Context in JavaScript - <hr>
## Description 📜
I'll be explaining briefly about Execution Context in JavaScript in the documentation.
<hr>
## Domain of Contribution 📊
- [x] JavaScript
<hr>
## Directory
`winter-of-contributing/Web_Development/Javascript/Topics/Execution_Context`
<hr>
## Formats
- [x] Document
<hr>
I am a GWOC'21 contributor from Batch 1 of JavaScript. Please assign this issue to me. I assure you that this issue's title will be the same as my PR's title as well as I will be Contributing under the Guidelines and Code of Conduct.
|
non_process
|
js execution context in javascript description 📜 i ll be explaining briefly about execution context in javascript in the documentation domain of contribution 📊 javascript directory winter of contributing web development javascript topics execution context formats document i am a gwoc contributor from batch of javascript please assign this issue to me i assure you that this issue s title will be the same as my pr s title as well as i will be contributing under the guidelines and code of conduct
| 0
|
9,173
| 12,225,909,990
|
IssuesEvent
|
2020-05-03 08:13:41
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
missing parent
|
multi-species process
|
GO:0140403 should be a child of GO:0052170
(In case it isn't in another ticket, I couldn't see it)
|
1.0
|
missing parent -
GO:0140403 should be a child of GO:0052170
(In case it isn't in another ticket, I couldn't see it)
|
process
|
missing parent go should be a child of go in case it isn t in another ticket i couldn t see it
| 1
|
20,732
| 14,118,320,385
|
IssuesEvent
|
2020-11-08 13:10:06
|
allo-/firefox-profilemaker
|
https://api.github.com/repos/allo-/firefox-profilemaker
|
opened
|
Add "breaks" meta data
|
Infrastructure
|
The idea with breaks metadata is to have something like
```
[
{
"name": "indexed_db",
"type": "boolean",
"initial": false,
"label": "Disable IndexedDB",
"help_text": "(...)",
"addons": [],
"config": {
"dom.indexedDB.enabled": false
},
"breaks": {
"description": "IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data."
"type": "websites" # websites, (firefox) features, etc.
"what": [
"All sites, which require indexedDB to work",
"Locally saving data on sites, like for example save games or chatlogs",
"Example: Twitter now requires IndexedDB to work"
]}
}
]
```
Rendered as
```
Disable IndexedDB
(description)
[
**This setting can break some websites.**
IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data.
- All sites, which require indexedDB to work
- Locally saving data on sites, like for example save games or chatlogs
- Example: Twitter now requires IndexedDB to work
]
```
Instead of just having `Disable IndexedDB (breaks things)`.
Maybe this could have a short description and a long description and the bullet points in some collapsible element.
|
1.0
|
Add "breaks" meta data - The idea with breaks metadata is to have something like
```
[
{
"name": "indexed_db",
"type": "boolean",
"initial": false,
"label": "Disable IndexedDB",
"help_text": "(...)",
"addons": [],
"config": {
"dom.indexedDB.enabled": false
},
"breaks": {
"description": "IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data."
"type": "websites" # websites, (firefox) features, etc.
"what": [
"All sites, which require indexedDB to work",
"Locally saving data on sites, like for example save games or chatlogs",
"Example: Twitter now requires IndexedDB to work"
]}
}
]
```
Rendered as
```
Disable IndexedDB
(description)
[
**This setting can break some websites.**
IndexedDB is needed to store indexed data on your PC. This could for example be a save game or a chatlog. Disabling it will not only prevent websites from storing unique identifiers ('super cookies') but also from storing useful data.
- All sites, which require indexedDB to work
- Locally saving data on sites, like for example save games or chatlogs
- Example: Twitter now requires IndexedDB to work
]
```
Instead of just having `Disable IndexedDB (breaks things)`.
Maybe this could have a short description and a long description and the bullet points in some collapsible element.
|
non_process
|
add breaks meta data the idea with breaks metadata is to have something like name indexed db type boolean initial false label disable indexeddb help text addons config dom indexeddb enabled false breaks description indexeddb is needed to store indexed data on your pc this could for example be a save game or a chatlog disabling it will not only prevent websites from storing unique identifiers super cookies but also from storing useful data type websites websites firefox features etc what all sites which require indexeddb to work locally saving data on sites like for example save games or chatlogs example twitter now requires indexeddb to work rendered as disable indexeddb description this setting can break some websites indexeddb is needed to store indexed data on your pc this could for example be a save game or a chatlog disabling it will not only prevent websites from storing unique identifiers super cookies but also from storing useful data all sites which require indexeddb to work locally saving data on sites like for example save games or chatlogs example twitter now requires indexeddb to work instead of just having disable indexeddb breaks things maybe this could have a short description and a long description and the bullet points in some collapsible element
| 0
|
4,580
| 7,208,714,287
|
IssuesEvent
|
2018-02-07 04:53:06
|
rust-lang-nursery/futures-rs
|
https://api.github.com/repos/rust-lang-nursery/futures-rs
|
closed
|
Add `Task::park_timeout`
|
0.1-incompatible 0.2-blocker C-feature-request
|
I've been rewriting my spmc-broadcast library [`bus`](https://github.com/jonhoo/bus) to expose a `Futures`-based interface rather than `recv/try_recv` (see the [`futures` branch](https://github.com/jonhoo/bus/tree/futures)), and am running into an issue. Due to the way the library works internally, there exists a condition under which the writer won't know that it needs to `unpark` the readers when it does a write. If this happens on the write that fills the buffer, the writer is not blocked on writing further, and the readers aren't aware that they need to start reading.
In the non-futures version, this is "fixed" by parking the readers using `thread::park_timeout`. It's not elegant, but it avoids the otherwise costly solution necessary to fix the underlying issue. Unfortunately, with `Futures`, only `task::park` is available, leading to a deadlock in the library whenever the scenario above occurs.
Would adding `task::park_timeout` be possible in `Futures`? Are there any major drawbacks to doing so?
|
True
|
Add `Task::park_timeout` - I've been rewriting my spmc-broadcast library [`bus`](https://github.com/jonhoo/bus) to expose a `Futures`-based interface rather than `recv/try_recv` (see the [`futures` branch](https://github.com/jonhoo/bus/tree/futures)), and am running into an issue. Due to the way the library works internally, there exists a condition under which the writer won't know that it needs to `unpark` the readers when it does a write. If this happens on the write that fills the buffer, the writer is not blocked on writing further, and the readers aren't aware that they need to start reading.
In the non-futures version, this is "fixed" by parking the readers using `thread::park_timeout`. It's not elegant, but it avoids the otherwise costly solution necessary to fix the underlying issue. Unfortunately, with `Futures`, only `task::park` is available, leading to a deadlock in the library whenever the scenario above occurs.
Would adding `task::park_timeout` be possible in `Futures`? Are there any major drawbacks to doing so?
|
non_process
|
add task park timeout i ve been rewriting my spmc broadcast library to expose a futures based interface rather than recv try recv see the and am running into an issue due to the way the library works internally there exists a condition under which the writer won t know that it needs to unpark the readers when it does a write if this happens on the write that fills the buffer the writer is not blocked on writing further and the readers aren t aware that they need to start reading in the non futures version this is fixed by parking the readers using thread park timeout it s not elegant but it avoids the otherwise costly solution necessary to fix the underlying issue unfortunately with futures only task park is available leading to a deadlock in the library whenever the scenario above occurs would adding task park timeout be possible in futures are there any major drawbacks to doing so
| 0
|
585,449
| 17,497,706,134
|
IssuesEvent
|
2021-08-10 04:24:04
|
code-ready/crc
|
https://api.github.com/repos/code-ready/crc
|
closed
|
[API] Refactor the config to use HTTP verbs
|
priority/major kind/task
|
Currently the config endpoint uses `config/get`, `config/set` and `config/unset` to expose the functionality to deal with configuration. However, this seems to be a direct copy of the binary protocol and does not follow HTTP verbs as expected
|
1.0
|
[API] Refactor the config to use HTTP verbs - Currently the config endpoint uses `config/get`, `config/set` and `config/unset` to expose the functionality to deal with configuration. However, this seems to be a direct copy of the binary protocol and does not follow HTTP verbs as expected
|
non_process
|
refactor the config to use http verbs currently the config endpoint uses config get config set and config unset to expose the functionality to deal with configuration however this seems to be a direct copy of the binary protocol and does not follow http verbs as expected
| 0
|
12,177
| 14,741,968,154
|
IssuesEvent
|
2021-01-07 11:28:18
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Dynamic Draft Invoice - Late Fees
|
anc-process anp-1.5 ant-enhancement
|
In GitLab by @kdjstudios on Feb 28, 2019, 10:24
**Submitted by:** Kyle
**Helpdesk:** NA
**Server:** All
**Client/Site:** ALL
**Account:** ALL
**Issue:**
IN #1302 we found the possibility that when adjusting items on the draft invoice, the late fee is not updating dynamically. Below is the test case document which contains those scenarios.
`Existing Functionality of Late Fee`: [Existing Functionality of Late Fee](https://docs.google.com/spreadsheets/d/1Rx-6jyVIIUqEFDMxnP4Zwe5CIL8Vvcsz-XKGtshFbvg/edit?usp=sharing)
`Draft Invoice Test Cases`: [Draft Invoice Test Cases](https://docs.google.com/spreadsheets/d/13pJPni78oybFgp2wN69gSSj7eFGerI6b6m50p8mwn7Q/edit?usp=sharing)
|
1.0
|
Dynamic Draft Invoice - Late Fees - In GitLab by @kdjstudios on Feb 28, 2019, 10:24
**Submitted by:** Kyle
**Helpdesk:** NA
**Server:** All
**Client/Site:** ALL
**Account:** ALL
**Issue:**
IN #1302 we found the possibility that when adjusting items on the draft invoice, the late fee is not updating dynamically. Below is the test case document which contains those scenarios.
`Existing Functionality of Late Fee`: [Existing Functionality of Late Fee](https://docs.google.com/spreadsheets/d/1Rx-6jyVIIUqEFDMxnP4Zwe5CIL8Vvcsz-XKGtshFbvg/edit?usp=sharing)
`Draft Invoice Test Cases`: [Draft Invoice Test Cases](https://docs.google.com/spreadsheets/d/13pJPni78oybFgp2wN69gSSj7eFGerI6b6m50p8mwn7Q/edit?usp=sharing)
|
process
|
dynamic draft invoice late fees in gitlab by kdjstudios on feb submitted by kyle helpdesk na server all client site all account all issue in we found the possibility that when adjusting items on the draft invoice the late fee is not updating dynamically below is the test case document which contains those scenarios existing functionality of late fee draft invoice test cases
| 1
|
13,161
| 15,590,007,544
|
IssuesEvent
|
2021-03-18 08:49:40
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
prisma migrate gives: Error querying the database: db error: ERROR: syntax error at or near "NOT"
|
bug/1-repro-available kind/bug process/candidate team/migrations topic: migrate topic: postgresql
|
## Bug description
`prisma migrate dev --preview-feature` is failing with _Error querying the database: db error: ERROR: syntax error at or near "NOT"_
The full debug output is:
```
yarn prisma migrate dev --preview-feature
yarn run v1.22.10
warning ../package.json: No license field
$ /home/ec2-user/environment/node_modules/.bin/prisma migrate dev --preview-feature
prisma:loadEnv project root found at /home/ec2-user/environment/package.json +0ms
prisma:tryLoadEnv Environment variables loaded from /home/ec2-user/environment/.env +0ms
Environment variables loaded from .env
Prisma schema loaded from prisma/schema.prisma
Datasource "db": PostgreSQL database "ebdb", schema "public" at "localhost:5432"
prisma:migrateEngine:rpc starting migration engine with binary: /home/ec2-user/environment/node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x +0ms
prisma:migrateEngine:rpc SENDING RPC CALL {"id":1,"jsonrpc":"2.0","method":"devDiagnostic","params":{"migrationsDirectoryPath":"/home/ec2-user/environment/prisma/migrations"}} +5ms
prisma:migrateEngine:stderr Mar 09 12:00:17.017 INFO migration_engine: Starting migration engine RPC server git_hash="da6fafb57b24e0b61ca20960c64e2d41f9e8cff1" +0ms
prisma:migrateEngine:stderr Mar 09 12:00:17.031 INFO quaint::single: Starting a postgresql connection. +13ms
prisma:migrateEngine:stderr Mar 09 12:00:17.273 INFO DevDiagnostic:calculate_drift:sql_schema_from_migration_history: quaint::single: Starting a postgresql connection. +242ms
prisma:migrateEngine:rpc { jsonrpc: '2.0',
prisma:migrateEngine:rpc error:
prisma:migrateEngine:rpc { code: 4466,
prisma:migrateEngine:rpc message: 'An error happened. Check the data field for details.',
prisma:migrateEngine:rpc data:
prisma:migrateEngine:rpc { is_panic: false,
prisma:migrateEngine:rpc message:
prisma:migrateEngine:rpc 'Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT"\n 0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history\n at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270\n 1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift\n at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40\n 2: migration_core::api::DevDiagnostic\n at migration-engine/core/src/api.rs:106',
prisma:migrateEngine:rpc backtrace: null } },
prisma:migrateEngine:rpc id: 1 } +370ms
Error: Error: Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT"
0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history
at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270
1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift
at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40
2: migration_core::api::DevDiagnostic
at migration-engine/core/src/api.rs:106
at Object.registerCallback (/home/ec2-user/environment/node_modules/prisma/build/index.js:55009:26)
at MigrateEngine.handleResponse (/home/ec2-user/environment/node_modules/prisma/build/index.js:54882:38)
at LineStream.byline_12.default.on (/home/ec2-user/environment/node_modules/prisma/build/index.js:54967:18)
at LineStream.emit (events.js:198:13)
at LineStream.EventEmitter.emit (domain.js:448:20)
at addChunk (_stream_readable.js:288:12)
at readableAddChunk (_stream_readable.js:269:11)
at LineStream.Readable.push (_stream_readable.js:224:10)
at LineStream.Transform.push (_stream_transform.js:151:32)
at LineStream._pushBuffer (/home/ec2-user/environment/node_modules/prisma/build/index.js:54719:19)
error Command failed with exit code 1.
```
## How to reproduce
I am not entirely sure what the exact steps that lead to the problem are, but hopefully the following is helpful:
I had deleted the previous migrations folder by accident, which had a single migration file in it. This migration file set up a basic User model.
After seeing the above errors, I did try a complete refresh of the database (dropped and recreated), as well as uninstall/reinstall and reset of prisma. Unfortunately the problem persisted.
## Expected behavior
`prisma migrate` to run normally and create the migration sql.
## Prisma information
My current prisma schema is:
```
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
binaryTargets = ["native"]
}
model Place {
id Int @id @default(autoincrement())
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
place_id String @unique
users UserPlace[]
}
model User {
id Int @id @default(autoincrement())
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
email String @unique
firstName String
lastName String
promotion Boolean @default(false)
places UserPlace[]
}
model UserPlace {
id Int @id @default(autoincrement())
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
placeId Int?
place Place? @relation(fields: [placeId], references: [id])
userId Int?
user User? @relation(fields: [userId], references: [id])
}
```
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: Amazon Linux 2
- Database: PostgreSQL (localhost)
- Node.js version: v10.23.0
- Prisma version
```
prisma : 2.18.0
@prisma/client : 2.18.0
Current platform : rhel-openssl-1.1.x
Query Engine : query-engine da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/query-engine-rhel-openssl-1.1.x)
Migration Engine : migration-engine-cli da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x)
Introspection Engine : introspection-core da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/introspection-engine-rhel-openssl-1.1.x)
Format Binary : prisma-fmt da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/prisma-fmt-rhel-openssl-1.1.x)
Studio : 0.356.0
```
Thanks for your time!
|
1.0
|
prisma migrate gives: Error querying the database: db error: ERROR: syntax error at or near "NOT" - ## Bug description
`prisma migrate dev --preview-feature` is failing with _Error querying the database: db error: ERROR: syntax error at or near "NOT"_
The full debug output is:
```
yarn prisma migrate dev --preview-feature
yarn run v1.22.10
warning ../package.json: No license field
$ /home/ec2-user/environment/node_modules/.bin/prisma migrate dev --preview-feature
prisma:loadEnv project root found at /home/ec2-user/environment/package.json +0ms
prisma:tryLoadEnv Environment variables loaded from /home/ec2-user/environment/.env +0ms
Environment variables loaded from .env
Prisma schema loaded from prisma/schema.prisma
Datasource "db": PostgreSQL database "ebdb", schema "public" at "localhost:5432"
prisma:migrateEngine:rpc starting migration engine with binary: /home/ec2-user/environment/node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x +0ms
prisma:migrateEngine:rpc SENDING RPC CALL {"id":1,"jsonrpc":"2.0","method":"devDiagnostic","params":{"migrationsDirectoryPath":"/home/ec2-user/environment/prisma/migrations"}} +5ms
prisma:migrateEngine:stderr Mar 09 12:00:17.017 INFO migration_engine: Starting migration engine RPC server git_hash="da6fafb57b24e0b61ca20960c64e2d41f9e8cff1" +0ms
prisma:migrateEngine:stderr Mar 09 12:00:17.031 INFO quaint::single: Starting a postgresql connection. +13ms
prisma:migrateEngine:stderr Mar 09 12:00:17.273 INFO DevDiagnostic:calculate_drift:sql_schema_from_migration_history: quaint::single: Starting a postgresql connection. +242ms
prisma:migrateEngine:rpc { jsonrpc: '2.0',
prisma:migrateEngine:rpc error:
prisma:migrateEngine:rpc { code: 4466,
prisma:migrateEngine:rpc message: 'An error happened. Check the data field for details.',
prisma:migrateEngine:rpc data:
prisma:migrateEngine:rpc { is_panic: false,
prisma:migrateEngine:rpc message:
prisma:migrateEngine:rpc 'Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT"\n 0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history\n at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270\n 1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift\n at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40\n 2: migration_core::api::DevDiagnostic\n at migration-engine/core/src/api.rs:106',
prisma:migrateEngine:rpc backtrace: null } },
prisma:migrateEngine:rpc id: 1 } +370ms
Error: Error: Database error: Error querying the database: db error: ERROR: syntax error at or near "NOT"
0: sql_migration_connector::flavour::postgres::sql_schema_from_migration_history
at migration-engine/connectors/sql-migration-connector/src/flavour/postgres.rs:270
1: sql_migration_connector::sql_database_migration_inferrer::calculate_drift
at migration-engine/connectors/sql-migration-connector/src/sql_database_migration_inferrer.rs:40
2: migration_core::api::DevDiagnostic
at migration-engine/core/src/api.rs:106
at Object.registerCallback (/home/ec2-user/environment/node_modules/prisma/build/index.js:55009:26)
at MigrateEngine.handleResponse (/home/ec2-user/environment/node_modules/prisma/build/index.js:54882:38)
at LineStream.byline_12.default.on (/home/ec2-user/environment/node_modules/prisma/build/index.js:54967:18)
at LineStream.emit (events.js:198:13)
at LineStream.EventEmitter.emit (domain.js:448:20)
at addChunk (_stream_readable.js:288:12)
at readableAddChunk (_stream_readable.js:269:11)
at LineStream.Readable.push (_stream_readable.js:224:10)
at LineStream.Transform.push (_stream_transform.js:151:32)
at LineStream._pushBuffer (/home/ec2-user/environment/node_modules/prisma/build/index.js:54719:19)
error Command failed with exit code 1.
```
## How to reproduce
I am not entirely sure what the exact steps that lead to the problem are, but hopefully the following is helpful:
I had deleted the previous migrations folder by accident, which had a single migration file in it. This migration file set up a basic User model.
After seeing the above errors, I did try a complete refresh of the database (dropped and recreated), as well as uninstall/reinstall and reset of prisma. Unfortunately the problem persisted.
## Expected behavior
`prisma migrate` to run normally and create the migration sql.
## Prisma information
My current prisma schema is:
```
datasource db {
provider = "postgresql"
url = env("DATABASE_URL")
}
generator client {
provider = "prisma-client-js"
binaryTargets = ["native"]
}
model Place {
id Int @id @default(autoincrement())
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
place_id String @unique
users UserPlace[]
}
model User {
id Int @id @default(autoincrement())
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
email String @unique
firstName String
lastName String
promotion Boolean @default(false)
places UserPlace[]
}
model UserPlace {
id Int @id @default(autoincrement())
createdAt DateTime @default(now())
updatedAt DateTime @updatedAt
placeId Int?
place Place? @relation(fields: [placeId], references: [id])
userId Int?
user User? @relation(fields: [userId], references: [id])
}
```
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: Amazon Linux 2
- Database: PostgreSQL (localhost)
- Node.js version: v10.23.0
- Prisma version
```
prisma : 2.18.0
@prisma/client : 2.18.0
Current platform : rhel-openssl-1.1.x
Query Engine : query-engine da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/query-engine-rhel-openssl-1.1.x)
Migration Engine : migration-engine-cli da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/migration-engine-rhel-openssl-1.1.x)
Introspection Engine : introspection-core da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/introspection-engine-rhel-openssl-1.1.x)
Format Binary : prisma-fmt da6fafb57b24e0b61ca20960c64e2d41f9e8cff1 (at node_modules/@prisma/engines/prisma-fmt-rhel-openssl-1.1.x)
Studio : 0.356.0
```
Thanks for your time!
|
process
|
prisma migrate gives error querying the database db error error syntax error at or near not bug description prisma migrate dev preview feature is failing with error querying the database db error error syntax error at or near not the full debug output is yarn prisma migrate dev preview feature yarn run warning package json no license field home user environment node modules bin prisma migrate dev preview feature prisma loadenv project root found at home user environment package json prisma tryloadenv environment variables loaded from home user environment env environment variables loaded from env prisma schema loaded from prisma schema prisma datasource db postgresql database ebdb schema public at localhost prisma migrateengine rpc starting migration engine with binary home user environment node modules prisma engines migration engine rhel openssl x prisma migrateengine rpc sending rpc call id jsonrpc method devdiagnostic params migrationsdirectorypath home user environment prisma migrations prisma migrateengine stderr mar info migration engine starting migration engine rpc server git hash prisma migrateengine stderr mar info quaint single starting a postgresql connection prisma migrateengine stderr mar info devdiagnostic calculate drift sql schema from migration history quaint single starting a postgresql connection prisma migrateengine rpc jsonrpc prisma migrateengine rpc error prisma migrateengine rpc code prisma migrateengine rpc message an error happened check the data field for details prisma migrateengine rpc data prisma migrateengine rpc is panic false prisma migrateengine rpc message prisma migrateengine rpc database error error querying the database db error error syntax error at or near not n sql migration connector flavour postgres sql schema from migration history n at migration engine connectors sql migration connector src flavour postgres rs n sql migration connector sql database migration inferrer calculate drift n at migration engine connectors sql migration connector src sql database migration inferrer rs n migration core api devdiagnostic n at migration engine core src api rs prisma migrateengine rpc backtrace null prisma migrateengine rpc id error error database error error querying the database db error error syntax error at or near not sql migration connector flavour postgres sql schema from migration history at migration engine connectors sql migration connector src flavour postgres rs sql migration connector sql database migration inferrer calculate drift at migration engine connectors sql migration connector src sql database migration inferrer rs migration core api devdiagnostic at migration engine core src api rs at object registercallback home user environment node modules prisma build index js at migrateengine handleresponse home user environment node modules prisma build index js at linestream byline default on home user environment node modules prisma build index js at linestream emit events js at linestream eventemitter emit domain js at addchunk stream readable js at readableaddchunk stream readable js at linestream readable push stream readable js at linestream transform push stream transform js at linestream pushbuffer home user environment node modules prisma build index js error command failed with exit code how to reproduce i am not entirely sure what the exact steps that lead to the problem are but hopefully the following is helpful i had deleted the previous migrations folder by accident which had a single migration file in it this migration file set up a basic user model after seeing the above errors i did try a complete refresh of the database dropped and recreated as well as uninstall reinstall and reset of prisma unfortunately the problem persisted expected behavior prisma migrate to run normally and create the migration sql prisma information my current prisma schema is datasource db provider postgresql url env database url generator client provider prisma client js binarytargets model place id int id default autoincrement createdat datetime default now updatedat datetime updatedat place id string unique users userplace model user id int id default autoincrement createdat datetime default now updatedat datetime updatedat email string unique firstname string lastname string promotion boolean default false places userplace model userplace id int id default autoincrement createdat datetime default now updatedat datetime updatedat placeid int place place relation fields references userid int user user relation fields references environment setup os amazon linux database postgresql localhost node js version prisma version prisma prisma client current platform rhel openssl x query engine query engine at node modules prisma engines query engine rhel openssl x migration engine migration engine cli at node modules prisma engines migration engine rhel openssl x introspection engine introspection core at node modules prisma engines introspection engine rhel openssl x format binary prisma fmt at node modules prisma engines prisma fmt rhel openssl x studio thanks for your time
| 1
|
20,051
| 26,540,017,103
|
IssuesEvent
|
2023-01-19 18:27:29
|
nion-software/nionswift
|
https://api.github.com/repos/nion-software/nionswift
|
opened
|
Add ability for plug-in packages to register custom data structures (entities)
|
type - enhancement stage - planning level - difficult f - processing f - API f - plugins feature type - internal maintenance
|
Related to-do items:
- [ ] include schema during file writes; required to allow reference counting
- [ ] automatic version compatibility with schemas; know whether a schema is compatible with an older version
- [ ] fully qualified type names; instead of context sensitive entity types which are difficult to implement
- [ ] reference count non-root items (i.e. an interval shared by EELS and a graphic)
|
1.0
|
Add ability for plug-in packages to register custom data structures (entities) - Related to-do items:
- [ ] include schema during file writes; required to allow reference counting
- [ ] automatic version compatibility with schemas; know whether a schema is compatible with an older version
- [ ] fully qualified type names; instead of context sensitive entity types which are difficult to implement
- [ ] reference count non-root items (i.e. an interval shared by EELS and a graphic)
|
process
|
add ability for plug in packages to register custom data structures entities related to do items include schema during file writes required to allow reference counting automatic version compatibility with schemas know whether a schema is compatible with an older version fully qualified type names instead of context sensitive entity types which are difficult to implement reference count non root items i e an interval shared by eels and a graphic
| 1
|
385,398
| 11,420,271,319
|
IssuesEvent
|
2020-02-03 09:47:27
|
nhn/tui.grid
|
https://api.github.com/repos/nhn/tui.grid
|
closed
|
vue-grid 4.x 에서 서버 연동
|
4.x Enhancement Priority: Normal Question
|
<!--
To make it easier for us to help you, please include as much useful information as possible.
Useful Links:
- tutorial: https://github.com/nhn/tui.grid/tree/master/docs
- API/Example: https://nhn.github.io/tui.grid/latest
Before opening a new issue, please search existing issues https://github.com/nhn/tui.grid/issues
-->
**Summary**
4.x에서 서버연동을 해보고 있습니다.
아래는 현재 서버 연동 시 리턴 규격입니다.
[https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md](https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md) 기준
```
{
"result": true,
"data": {
"contents": [],
"pagination": {
"page": 1,
"totalCount": 100
}
}
}
```
공통 규격이 있는 상태에서 그리드 용도의 api만 규격에 맞게 하기는 비효율인 것 같아서요,
위 규격이 아닌 커스텀한 데이터를 할당할 수 있는지요?
ex) 아래 규격 중 data 를 할당
```
{
code: 200,
message: "정상 처리 되었습니다",
data: { result: true, data: { contents: ......, pagination: { ....... } } }
}
```
**Screenshots**
If applicable, add screenshots to help explain your question.
**Version**
toast-ui.vue-grid 2.0.1
**Additional context**
추가로 문의드립니다.
1. 서버사이드 구성 시 data값을 아래와 같이 했더니 데이터는 잘 나옵니다만
Invalid prop: type check failed for prop "data". Expected Array, got Object 라는 오류가 발생합니다.
어떻게 할당을 해야 하는지요?
2. 페이지 구성이 되지 않고 있습니다. 어디가 잘못된 건지 알려주시면 감사하겠습니다 ㅠ.ㅠ
```
<template>
<grid ref="tuiGrid" :data="gridProps.data" :columns="gridProps.columns" :options="gridProps.options"/>
</template>
<script>
import 'tui-grid/dist/tui-grid.css'
import { Grid } from '@toast-ui/vue-grid'
export default {
components: {
'grid': Grid
},
created() {
this.gridProps = {
data: {
api: {
readData: { url: '/api/grid/test', method: 'GET' },
}
},
pageOptions: {
perPage: 5
},
pagination: true,
columns: [
{ name: "name", header: "이름", align: "center", sortable: true },
{ name: "age", header: "나이", align: "center", sortable: true },
{ name: "addr", header: "주소", align: "center" }
]
}
},
methods: {
}
}
</script>
<style>
</style>
```
|
1.0
|
vue-grid 4.x 에서 서버 연동 - <!--
To make it easier for us to help you, please include as much useful information as possible.
Useful Links:
- tutorial: https://github.com/nhn/tui.grid/tree/master/docs
- API/Example: https://nhn.github.io/tui.grid/latest
Before opening a new issue, please search existing issues https://github.com/nhn/tui.grid/issues
-->
**Summary**
4.x에서 서버연동을 해보고 있습니다.
아래는 현재 서버 연동 시 리턴 규격입니다.
[https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md](https://github.com/nhn/tui.grid/blob/master/docs/ko/data-source.md) 기준
```
{
"result": true,
"data": {
"contents": [],
"pagination": {
"page": 1,
"totalCount": 100
}
}
}
```
공통 규격이 있는 상태에서 그리드 용도의 api만 규격에 맞게 하기는 비효율인 것 같아서요,
위 규격이 아닌 커스텀한 데이터를 할당할 수 있는지요?
ex) 아래 규격 중 data 를 할당
```
{
code: 200,
message: "정상 처리 되었습니다",
data: { result: true, data: { contents: ......, pagination: { ....... } } }
}
```
**Screenshots**
If applicable, add screenshots to help explain your question.
**Version**
toast-ui.vue-grid 2.0.1
**Additional context**
추가로 문의드립니다.
1. 서버사이드 구성 시 data값을 아래와 같이 했더니 데이터는 잘 나옵니다만
Invalid prop: type check failed for prop "data". Expected Array, got Object 라는 오류가 발생합니다.
어떻게 할당을 해야 하는지요?
2. 페이지 구성이 되지 않고 있습니다. 어디가 잘못된 건지 알려주시면 감사하겠습니다 ㅠ.ㅠ
```
<template>
<grid ref="tuiGrid" :data="gridProps.data" :columns="gridProps.columns" :options="gridProps.options"/>
</template>
<script>
import 'tui-grid/dist/tui-grid.css'
import { Grid } from '@toast-ui/vue-grid'
export default {
components: {
'grid': Grid
},
created() {
this.gridProps = {
data: {
api: {
readData: { url: '/api/grid/test', method: 'GET' },
}
},
pageOptions: {
perPage: 5
},
pagination: true,
columns: [
{ name: "name", header: "이름", align: "center", sortable: true },
{ name: "age", header: "나이", align: "center", sortable: true },
{ name: "addr", header: "주소", align: "center" }
]
}
},
methods: {
}
}
</script>
<style>
</style>
```
|
non_process
|
vue grid x 에서 서버 연동 to make it easier for us to help you please include as much useful information as possible useful links tutorial api example before opening a new issue please search existing issues summary x에서 서버연동을 해보고 있습니다 아래는 현재 서버 연동 시 리턴 규격입니다 기준 result true data contents pagination page totalcount 공통 규격이 있는 상태에서 그리드 용도의 api만 규격에 맞게 하기는 비효율인 것 같아서요 위 규격이 아닌 커스텀한 데이터를 할당할 수 있는지요 ex 아래 규격 중 data 를 할당 code message 정상 처리 되었습니다 data result true data contents pagination screenshots if applicable add screenshots to help explain your question version toast ui vue grid additional context 추가로 문의드립니다 서버사이드 구성 시 data값을 아래와 같이 했더니 데이터는 잘 나옵니다만 invalid prop type check failed for prop data expected array got object 라는 오류가 발생합니다 어떻게 할당을 해야 하는지요 페이지 구성이 되지 않고 있습니다 어디가 잘못된 건지 알려주시면 감사하겠습니다 ㅠ ㅠ import tui grid dist tui grid css import grid from toast ui vue grid export default components grid grid created this gridprops data api readdata url api grid test method get pageoptions perpage pagination true columns name name header 이름 align center sortable true name age header 나이 align center sortable true name addr header 주소 align center methods
| 0
|
630,658
| 20,115,963,097
|
IssuesEvent
|
2022-02-07 19:32:51
|
IDAES/examples-pse
|
https://api.github.com/repos/IDAES/examples-pse
|
closed
|
More failures in example notebooks
|
bug Priority:High
|
The recent merge of https://github.com/IDAES/idaes-pse/pull/627 has revealed additional issues with the HDS flowsheet example (Examples/Flowsheets/HDA_flowsheet_with_distillation_solution_testing.ipynb). It appears the reactor initialization step is/was failing and that it had not been noticed up until now - now that the CSTR initialization routine raises an `InitializationError` if the final solve fails to converge this notebook is now failing.
We need to look at this model once again and determine what the issue is and try to improve the models performance (hopefully through scaling). As an absolute last resort, we can wrap the CSTR initialization step in a `try/except` statement to skip the error, but that should be avoided as this is our flagship tutorial example.
|
1.0
|
More failures in example notebooks - The recent merge of https://github.com/IDAES/idaes-pse/pull/627 has revealed additional issues with the HDS flowsheet example (Examples/Flowsheets/HDA_flowsheet_with_distillation_solution_testing.ipynb). It appears the reactor initialization step is/was failing and that it had not been noticed up until now - now that the CSTR initialization routine raises an `InitializationError` if the final solve fails to converge this notebook is now failing.
We need to look at this model once again and determine what the issue is and try to improve the models performance (hopefully through scaling). As an absolute last resort, we can wrap the CSTR initialization step in a `try/except` statement to skip the error, but that should be avoided as this is our flagship tutorial example.
|
non_process
|
more failures in example notebooks the recent merge of has revealed additional issues with the hds flowsheet example examples flowsheets hda flowsheet with distillation solution testing ipynb it appears the reactor initialization step is was failing and that it had not been noticed up until now now that the cstr initialization routine raises an initializationerror if the final solve fails to converge this notebook is now failing we need to look at this model once again and determine what the issue is and try to improve the models performance hopefully through scaling as an absolute last resort we can wrap the cstr initialization step in a try except statement to skip the error but that should be avoided as this is our flagship tutorial example
| 0
|
460,029
| 13,203,457,661
|
IssuesEvent
|
2020-08-14 14:12:24
|
novelis-prod/Digital-CoE-Operations-Data---Public
|
https://api.github.com/repos/novelis-prod/Digital-CoE-Operations-Data---Public
|
closed
|
Pinda tables not partitioned or contain more than just partitions within opsentprodg2
|
Plant: Pinda Priority #1
|
Went to look at which dates the pinda tables had within opsentprodg2 and found that many tables are no longer partitioned by anything, or contained partitions and parquet files outside of those partitions










|
1.0
|
Pinda tables not partitioned or contain more than just partitions within opsentprodg2 - Went to look at which dates the pinda tables had within opsentprodg2 and found that many tables are no longer partitioned by anything, or contained partitions and parquet files outside of those partitions










|
non_process
|
pinda tables not partitioned or contain more than just partitions within went to look at which dates the pinda tables had within and found that many tables are no longer partitioned by anything or contained partitions and parquet files outside of those partitions
| 0
|
189,129
| 22,046,987,266
|
IssuesEvent
|
2022-05-30 03:39:49
|
madhans23/linux-4.1.15
|
https://api.github.com/repos/madhans23/linux-4.1.15
|
closed
|
CVE-2016-6213 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed
|
security vulnerability
|
## CVE-2016-6213 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fs/namespace.c in the Linux kernel before 4.9 does not restrict how many mounts may exist in a mount namespace, which allows local users to cause a denial of service (memory consumption and deadlock) via MS_BIND mount system calls, as demonstrated by a loop that triggers exponential growth in the number of mounts.
<p>Publish Date: 2016-12-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6213>CVE-2016-6213</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6213">https://nvd.nist.gov/vuln/detail/CVE-2016-6213</a></p>
<p>Release Date: 2016-12-28</p>
<p>Fix Resolution: 4.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-6213 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2016-6213 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/pnode.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fs/namespace.c in the Linux kernel before 4.9 does not restrict how many mounts may exist in a mount namespace, which allows local users to cause a denial of service (memory consumption and deadlock) via MS_BIND mount system calls, as demonstrated by a loop that triggers exponential growth in the number of mounts.
<p>Publish Date: 2016-12-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-6213>CVE-2016-6213</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-6213">https://nvd.nist.gov/vuln/detail/CVE-2016-6213</a></p>
<p>Release Date: 2016-12-28</p>
<p>Fix Resolution: 4.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files fs pnode c vulnerability details fs namespace c in the linux kernel before does not restrict how many mounts may exist in a mount namespace which allows local users to cause a denial of service memory consumption and deadlock via ms bind mount system calls as demonstrated by a loop that triggers exponential growth in the number of mounts publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
175,760
| 6,553,706,095
|
IssuesEvent
|
2017-09-06 00:23:53
|
opencurrents/opencurrents
|
https://api.github.com/repos/opencurrents/opencurrents
|
closed
|
upcoming-events: Events only displaying for admins of the org who created event
|
priority high
|
Volunteers are not seeing any events
|
1.0
|
upcoming-events: Events only displaying for admins of the org who created event - Volunteers are not seeing any events
|
non_process
|
upcoming events events only displaying for admins of the org who created event volunteers are not seeing any events
| 0
|
18,660
| 24,581,464,637
|
IssuesEvent
|
2022-10-13 15:55:36
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[FHIR] Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file
|
Bug P0 Response datastore Process: Fixed Process: Tested dev
|
AR: Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file
ER: Questionnaires resources > JSON > 'Title' key should get displayed in the JSON file
|
2.0
|
[FHIR] Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file - AR: Questionnaires resources > JSON > 'Title' key is not getting displayed in the JSON file
ER: Questionnaires resources > JSON > 'Title' key should get displayed in the JSON file
|
process
|
questionnaires resources json title key is not getting displayed in the json file ar questionnaires resources json title key is not getting displayed in the json file er questionnaires resources json title key should get displayed in the json file
| 1
|
114,119
| 14,531,533,997
|
IssuesEvent
|
2020-12-14 20:56:29
|
retaildevcrews/ngsa
|
https://api.github.com/repos/retaildevcrews/ngsa
|
closed
|
Review Data Separation Design
|
DesignReview PlatValidApp ValidDemo
|
## Description
What:
- Review and discuss design considerations from the data separation spike #168
Why:
- Drive understanding of new design
- Collect any feedback prior to incorporating into codebase
When:
- ASAP
Where:
- ngsa app (C#, TS?)
## Tasks
- [x] Discuss and Review design #197
- [ ] Document design #196
## Acceptance Criteria
- Meeting had and information delivered
- Any critical questions addressed
- Design documented
-
## Constraints
- N/A
## References:
- References #8
|
1.0
|
Review Data Separation Design - ## Description
What:
- Review and discuss design considerations from the data separation spike #168
Why:
- Drive understanding of new design
- Collect any feedback prior to incorporating into codebase
When:
- ASAP
Where:
- ngsa app (C#, TS?)
## Tasks
- [x] Discuss and Review design #197
- [ ] Document design #196
## Acceptance Criteria
- Meeting had and information delivered
- Any critical questions addressed
- Design documented
-
## Constraints
- N/A
## References:
- References #8
|
non_process
|
review data separation design description what review and discuss design considerations from the data separation spike why drive understanding of new design collect any feedback prior to incorporating into codebase when asap where ngsa app c ts tasks discuss and review design document design acceptance criteria meeting had and information delivered any critical questions addressed design documented constraints n a references references
| 0
|
446,472
| 31,478,604,463
|
IssuesEvent
|
2023-08-30 12:30:26
|
TycheSoftwares/woocommerce-abandoned-cart
|
https://api.github.com/repos/TycheSoftwares/woocommerce-abandoned-cart
|
closed
|
Update the screenshots in the documentation.
|
type: update documentation
|
**Describe the bug**
Update the screenshots in the documentation as per the current UI of tabs.
https://www.tychesoftwares.com/docs/docs/abandoned-cart-for-woocommerce-lite/capturing-abandoned-carts/
|
1.0
|
Update the screenshots in the documentation. - **Describe the bug**
Update the screenshots in the documentation as per the current UI of tabs.
https://www.tychesoftwares.com/docs/docs/abandoned-cart-for-woocommerce-lite/capturing-abandoned-carts/
|
non_process
|
update the screenshots in the documentation describe the bug update the screenshots in the documentation as per the current ui of tabs
| 0
|
16,858
| 22,139,784,934
|
IssuesEvent
|
2022-06-03 05:06:53
|
arcus-azure/arcus.messaging
|
https://api.github.com/repos/arcus-azure/arcus.messaging
|
closed
|
Remove reflection from general and Azure Service Bus-specific message routing
|
area:message-processing
|
### Discussed in https://github.com/arcus-azure/arcus.messaging/discussions/261
<div type='discussions-op-text'>
<sup>Originally posted by **fgheysels** March 24, 2022</sup>
Not an idea for a new feature, but an idea for a possible improvement.
As mentionned in this PR #260 , the code to retrieve registered `IMessageHandler` implementations is not that stable. We should find a way to improve this.
@stijnmoreels's comment copied from the mentionned PR:
> You'll probably wonder why we have this reflection code in the MessageHandler file.
This is bc we were faced with a problem when implementing the message routing. The IServceProvider doesn't support open generic interfaces and since our IMessageHandler<,> is such an open interface, there is no build-in way to extract all service registrations with a generic message type.
This message handler class will do this for us. It searches for the 'engine' and 'call site factory' on the service provider and extracts all the message handler interfaces. Later on, we can determine if there exists and message handler that can process the incoming message.
>
>We can off course discuss this later, or come up with something else. For now, this was the best we could do. Unfortunately.
I was wondering if we couldn't work around this without using reflection.
Maybe it is an idea to introduce a non-generic interface (empty) `IMessageHandler`. The `IMessageHandler<,>` would then derive from that non-generic `IMessageHandler` interface, like this:
```csharp
public interface IMessageHandler
{
}
public interface IMessageHandler<in TMessage, in TMessageContext> : IMessageHandler where TMessageContext : MessageContext
{
...
}
```
The code for retrieving registered message-handlers could then work on retrieving types that are assignable from the base-interface `IMessageHandler`.
Once we have retrieved those types, then we need of course an additional filter: we only want to take into consideration those types that are also assignable to the IMessageHandler interface which declares type parameters.</div>
|
1.0
|
Remove reflection from general and Azure Service Bus-specific message routing - ### Discussed in https://github.com/arcus-azure/arcus.messaging/discussions/261
<div type='discussions-op-text'>
<sup>Originally posted by **fgheysels** March 24, 2022</sup>
Not an idea for a new feature, but an idea for a possible improvement.
As mentionned in this PR #260 , the code to retrieve registered `IMessageHandler` implementations is not that stable. We should find a way to improve this.
@stijnmoreels's comment copied from the mentionned PR:
> You'll probably wonder why we have this reflection code in the MessageHandler file.
This is bc we were faced with a problem when implementing the message routing. The IServceProvider doesn't support open generic interfaces and since our IMessageHandler<,> is such an open interface, there is no build-in way to extract all service registrations with a generic message type.
This message handler class will do this for us. It searches for the 'engine' and 'call site factory' on the service provider and extracts all the message handler interfaces. Later on, we can determine if there exists and message handler that can process the incoming message.
>
>We can off course discuss this later, or come up with something else. For now, this was the best we could do. Unfortunately.
I was wondering if we couldn't work around this without using reflection.
Maybe it is an idea to introduce a non-generic interface (empty) `IMessageHandler`. The `IMessageHandler<,>` would then derive from that non-generic `IMessageHandler` interface, like this:
```csharp
public interface IMessageHandler
{
}
public interface IMessageHandler<in TMessage, in TMessageContext> : IMessageHandler where TMessageContext : MessageContext
{
...
}
```
The code for retrieving registered message-handlers could then work on retrieving types that are assignable from the base-interface `IMessageHandler`.
Once we have retrieved those types, then we need of course an additional filter: we only want to take into consideration those types that are also assignable to the IMessageHandler interface which declares type parameters.</div>
|
process
|
remove reflection from general and azure service bus specific message routing discussed in originally posted by fgheysels march not an idea for a new feature but an idea for a possible improvement as mentionned in this pr the code to retrieve registered imessagehandler implementations is not that stable we should find a way to improve this stijnmoreels s comment copied from the mentionned pr you ll probably wonder why we have this reflection code in the messagehandler file this is bc we were faced with a problem when implementing the message routing the iservceprovider doesn t support open generic interfaces and since our imessagehandler is such an open interface there is no build in way to extract all service registrations with a generic message type this message handler class will do this for us it searches for the engine and call site factory on the service provider and extracts all the message handler interfaces later on we can determine if there exists and message handler that can process the incoming message we can off course discuss this later or come up with something else for now this was the best we could do unfortunately i was wondering if we couldn t work around this without using reflection maybe it is an idea to introduce a non generic interface empty imessagehandler the imessagehandler would then derive from that non generic imessagehandler interface like this csharp public interface imessagehandler public interface imessagehandler imessagehandler where tmessagecontext messagecontext the code for retrieving registered message handlers could then work on retrieving types that are assignable from the base interface imessagehandler once we have retrieved those types then we need of course an additional filter we only want to take into consideration those types that are also assignable to the imessagehandler interface which declares type parameters
| 1
|
7,134
| 10,278,478,759
|
IssuesEvent
|
2019-08-25 14:45:57
|
nextmoov/nextmoov
|
https://api.github.com/repos/nextmoov/nextmoov
|
closed
|
Getting Started Dev Templates
|
#Dev Tools & Processes
|
-[] nodejs vanilla esNext
-[] nodejs vanilla esNext + express
-[] F-E vanilla esNext
-[] F-E vanilla esNext + react
-[] F-E vanilla esNext + react+ redux
|
1.0
|
Getting Started Dev Templates - -[] nodejs vanilla esNext
-[] nodejs vanilla esNext + express
-[] F-E vanilla esNext
-[] F-E vanilla esNext + react
-[] F-E vanilla esNext + react+ redux
|
process
|
getting started dev templates nodejs vanilla esnext nodejs vanilla esnext express f e vanilla esnext f e vanilla esnext react f e vanilla esnext react redux
| 1
|
747,221
| 26,077,585,092
|
IssuesEvent
|
2022-12-24 19:57:20
|
zigtools/zls
|
https://api.github.com/repos/zigtools/zls
|
closed
|
Another integer overflow in `fullPtrType`
|
bug priority:high fuzzing result
|
### Zig Version
0.11.0-dev.782+0b4461d97
### Zig Language Server Version
0.11.0-dev.72+3526f5f
### Steps to Reproduce
Open the following file:
```zig
**CapsuleHeader,
.description = "SPIR-V version = version.
var x69: u1 = undefined;
addcarryxU64(&x112, &x113, x110);
try std.testing.expectEqualSlices(u8, &bytes, &[_]u8{}, it.peek().?, "d");
length is less than `locked`.
/// For example, instead of file action type, but it is safe for user protocol = extern struct {
pub const CGT_SYM = "__vdso_clock_getres50(clk_id: c_int) isize;
pub extern "c" fn posix_spawnattr_setguardsize(attr: *posix_spawnattr_t = extern struct {
// Ensure that there must be freed before the window is not implementation failed with @asyncCall: HANDLE,
// TODO: Remove and return error.Permissions for Embedded ProcessForWsWatch(hProcess: [*]u8,
.NOMEM => return error.SystemResources,
const math = std.meta;
const ld_info.name));
} else |err| @as(i32, -1), @as(i32, 4), queue.removeMin(), 1);
/// with the state as bytes in POSIX message into memory copy, but you are connectionRefused;
}
pub fn proc_exit(0);
.allocator };
}
// Unimplementation information about to do it.
.tag = .decl_between_source };
}
/// Compute x*y (mod L)
pub const p: anyframe },
.TrueLiteral1;
alg_version) |ver| {
const sqe_recv = net.AddressInUse,
var u2a = U2{ .C = 0 };
proc: extern "user32" fn Dispatch" {
\\\\ switch (w.args_sizes_get_ret = os.wasi.clock_time_get_ns,
end_ix = original_deps[0].name, "A"));
try bit_stream.readBits(u5, 2, &out_bits)) {
else
extern union {
return self.unmanaged{ .integer startup key or signal_t = 1;
var off_out);
.fresh => log.info("FrameworkWeak(self: *Self, elem: T) []T {
t4 = t4.mul(Y3);
if (len == b.len) : (offset < try seekable,
var k: i32 = undefined behavior of non-instructions",
try expectFmt("hello world!", "{[1]s} world{[punctuation is correct type.
ATOMIC_LOCKS_NOT_STARTED = 0x00000004;
/// Get the pointer to these bits < byte_size;
}
/// Delete a direct writes if more than the anyframe->ReturnType(comptime buffer size (in bytes
const s_per_min;
pub const ELFDATA2MSB => .Big,
.dependencies = featureSet(&[_]Feature{
\\\\
);
}
test "y_object_id_t = i32;
pub const value: [2]u64,
/// therefore the equality here on, we only need to double free
&out_port,
// `new_len` must be called on this command is of type `T` has a bug",
break;
}
}
return self.unmanaged(Node.Index) full.FnProto {
var r = try std.testing.expect(mem.eql(u8, it.next()) {
\\\\};
const len = Codecs{
.android) struct {
: [number] "{rbp}" (@enumToInt(Feature.lvi_load_hardening",
field_state = .u8;
}
/// A system cannot be deleted.
RPC_S_OBJECT_PATH_SYNTAX_BAD => unreachable,
return error.EndOfStream {
inline data.
/// When trying next members_range.end],
.level_3,
return p.addNode(.{
}
.tail_call = 0,
while (i > 0) {
try testing.expectEqualSlices(u8, expected` == hex(`input`) where || is concatenation
fn parse(s: []const u8, dir_list: std.c.MACH_TASK_BASIC_INFORMATION = 0xC01E0354,
NOT_CONTAINER_STATE_VETOED = 0x8000000000000000000000000000000000000000000000000000000, 0x22, 0x68, 0x65, 0x6e, 0xe8, 0x90, 0x27, 0x3f, 0xdf, 0xfd616b15, 0x2f501ec8, 0xad0552ab,
0x07cfe89cf1bd76ac,
\\\\["\\u200B"]
);
}
test "y_array_expr,
r: []Limb) usize {
\\\\pub extern "c" fn thread_id: u64,
.{ .key = "042927b10512bae3eddcfe467828128bad2903269919f7086069c8c4df6c732838c7787964eaac00e5921fb1498a60f4606766b3d9685001558d1a974e7341513e", .msg = "313233343030", .sig = "3026090180022100b329f479a2bbd0a5c384ee1493b1f5186a87139cac5df4087c134b49156847db", .result = try os.openbsd.FUTEX_WAIT takes no requeue address + end_offset, shdr.sh_size) {
try testing.expect(isWhitespace.
FWP_DYNAMIC
.llvm_name = "arm1020e",
.step = Step.init(.run, name, ".debug_info = @typeInfo(T).Int.bits) return Literal e exponent -= exponent_fast_path) catch null;
E{ .part = 0xc07, .m32 = &A64.cortex_a8",
\\\\ # Return the_copy.dependencies = featureSet(&[_]Feature{
}
}
.rclass,
fp16fml,
// added the state byte_count = fields_delimiterArrayListUnmanaged.clearRetainingCapacityContext(key: [32]u8) Curve25519BasePointPc[0..9] else pc: {
else if (mem.eql(u8, name, ".")) {
.llvm_name = null,
single_threaded) {
.fast_skip_hashshing = 6,
ftruncate(u64, x >> 4));
mem.swap(Self, self.fd, sub_path_w, new_dir_mode);
.unsigned, math.log1p(x + x * r32(x * x);
if (i >= iovecs[i].iov_len = @sizeOf(c_int) aligned_addr == 0 and i & 2 != 0, a.isSet(0));
// present.
MUTUAL_AUTH_FAILED = 1634,
log2_buf_align: u8, ret_addr: ?*sockaddr, addr_size: *usize, value: T, endian);
another allocator: mem.Allocator);
.clwb,
try std.testing.allocator, u);
self.rd_pos..self.window_end == 0) try p.warnMsg(.{
.@"3e3r2",
return badDwarf();
/// Write `string` to `writer` is not valid.
INVALID_ICON_HANDLE = IOC_OUT | IOC_WS2 | 27;
pub const WM_CAP_FILE_SAVEASW = 0x0479;
pub const IOCOUTQ = if (is_array) |_, bit_index + 1;
\\\\ asm volatile ("svc #0"
.main_token = p.assertTuple(.{u32}, Tuple(&[_]type{ f16, f32, f64, f80, f128, c_long, 124);
h.update(options.allocator, 1);
.time_low = 0xa7af67cb,
handler support mirroring display mode in word width
pub const ABRT = 6;
/// but this extra buffer len
/// dropped a clear the base-2 shift. This swaps the literal = .llu };
}
}
CPU,
.description = "Output struct {
return self.endianSwap();
inline while (i < 16) : (j += 1) {
}
return FileSource,
state.end_index -= 1;
}
error.Utf8OverlongEncodingError.NoSpaceLeft,
.avr0,
addcarryxU64(&x79, &x80, x1, (x13 & 0xc7634d81f4372ddf);
if (index += 1;
.pku,
0xb0ba0da5bb600397,
sgetmask = 200, .strx = undefined;
subborrowxU64(&x57, &x58, x56, x19, x21);
}
One,
error.FileDescriptor could not used.
ABORTED = 0xC00000CA,
.INVAL => error.AccessDenied,
.float1e2,
try expectEqualStrings(self.bits.set(index)) != 0;
fbs.pos += n;
.Shader,
SPV_INTEL_unstructure pointer to the specified in this binary is foreign binaries.
/// Format = struct {
fd: os.fd_t,
/// The ArrayList(u8).init(key, npub);
},
return Self{
const b10 = (s0 -% y0) >> 63;
pub const IP_WFP_REDIRECT_CONTEXT_NOT_FOUND => return result.tag = .caret;
for (section) |only_section: std.os.iovec;
extern "wasi_snapshot_preview1" fn fd_filestat_set_times(fd: fd_t, buf: [*]u8,
fast_gather,
.btst16,
/// IPsec dropped when MH_PREBOUND_DYLIB = (0x18 | LC_REQ_DYLD),
\\\\ "tab": "\\t",
.clock_settime64 = 421,
PIPE_NOWAIT = 0xC00002B1,
f32x4_relaxed_laneselect = 142,
secs: u17, // max avail throughput });
'=' => {
\\\\ , & && &= ? < <= <<
.LOOP => return GzipStream(allocator, maxInt(u3))
}
.description = "Reserve X17",
}
const c = p.t.mul(z_inv); // T*z_inv
.fullfp16,
},
.Dynamic => try p.warn(.expected_continuation);
}
/// `queryCapsuleCapability StoragePushConstant(in_stream.readIntBig(u128, &l);
compiler of Zig, using this function subborrowxU64(&x69, &x70, x1, (arg2[3]));
try testing.expectEqual(linux.getuid());
const fd = @intCast(MinInt)`
},
tp.* = .{
// Note: can't splice
mem.writeIntLittle(u16, utf16le_buf.len, mem.page_size, shstr64.sh_offset).* += base_addr = @ptrCast(*FailingAlloc(alloc, "\\"foo\\\
bar\\")));
prev: ?*LinkMap,
xsavec,
inherit it
/// The checks are a
/// primitive integer.
WSA_E_CANCELLED = 0xC0220013,
/// multiple process argv/env
pub const b = Complex = cmath.Complex;
/// PrivateUsage: SIZE_T,
PUT,
\\\\}
IV[1],
},
if (args.len != 64) return self.resize(self.items[i], items[2]);
var kf = invln2 * x;
defer tmp_dir.dir.makePath(self.builder.dupe(phrase),
}
}
}
fn renderExpressions: ?[]const u8 = null,
defer a.deinit(self: FileSource,
}
};
const index_size = dyn_size = hasher.final();
}
fn renderToken(ais, tree, body, space);
.vector_enhancement for slice[read_idx] == '.') orelse return error.SystemResources,
if (!node.view.pending.fetchSub(5, ordering),
const window parameter to IndexedArray(EnumIndexer(Indexer);
defer std.testing;
const WINAPI or STDCALL.
/// {Missing System and return mem.eql(u8, content = true;
return @enumToInt(Feature.avx512f,
.Keyword_do,
Insn.ld_imm(0),
slab("24641172998046401", -56),
slab("2229658838863212", -132),
.NoReturn: void,
R9: DWORD64,
var x43: u64 = undefined;
cmovznzU64(&x216, x215, x212, x235);
var e = DeflateFast{
iu += 0x3F800000) {
return pcMul16(pc: *const fn (c_int) + 1, .space);
defer tmp.cleanup();
: [number], %%eax
while (true) {
try checkNext(&p, .String = .{ .StringLiteral(c_ulong) c_long, 50);
pub const POOL = 49;
pub const LUP_RETURN_ALIASES = 1024;
pub const bit = @intCast(ShelfIndex) usize {
.user_data = user_data: u32,
mips5_32r2,
v.* = Stack.init();
/// File exists and the password,
ring.deinit();
}
const ad = "Additional div" {
const x54 = (x42 + x53);
var buf: [8192]u8 = undefined;
var enabled, to run on Valgrind() usize {
writercv: pthread_create(name: *const [16 * count]u8 = undefined. Note: microsoft/microsoft.com/en-us/library/windows/ole32.zig");
pub const imported global_instance = @ptrCast(*anyopaque = blk: {
const bytes = [_:0]u16{0} ** codegen and codegen_code_count]u8 = undefined;
var i: usize) void {
}
.treap = Treap(u64, u64).init(r, y - y);
artifact.target.isDarwin(tag: Tag, arch: Cpu.Arch) Machine Scheduling policy failure = .{ .path = "test_file.txt");
h = Blake2b(Box.nonce_length]u8 = [_]u8{};
defer tmp.cleanup();
const fn (context, lessThan(i32, 0, 5) catch return null;
}
// FileTime(*FILETIME,
lpWSAData: *WSABUF,
dwProviderFlags,
/// The new Mutable and this map
&out_port,
\\\\push $0
}
.avoid_partial;
builder.pathJoin(&lib_list[0] = p[0];
const iovec = std.os.windows) {
.next = undefined;
cmovznzU64(&x295, &x296, x258, x255);
var buf1: [1]u8 = undefined;
mulxU64(&x331, &x332, x5, 0xffffffffffffff);
var acc = st.atime(),
else => {},
});
\\\\}
return q;
}
return readlinkW(file_path);
var t2: ?TokenIndex, space: Space = enum(u32) {
p.count = 0,
destination addcarryxU64(&x47, &x48, x1, (arg2[2]), (arg3[2]), x18);
var result: [len]CpuFeature.ccdp)] = .{
.disk_designatorWindows("\\\\\\\\unc\\\\share\\\\foo\\\\", "\\\\\\\\foo\\\\bar", "c:\\\\foo\\\\tmp.3\\\\", "..\\\\baz");
if (name.len);
}
0x30, 0x2f83071f53f325dd,
f32_store = 0x38,
flags: u32) c_long, 76);
pub const s1 = try toPosixPath(sub_dir_path);
if (callconv(WINAPI) NTSTATUS;
pub extern "user32" fn GetCurrentDirectory as the index in `entries` must be used instead"),
);
try testing.expect(isSorted(u8, "zyxw", {}, desc_u8));
defer q.deinit();
attr.map_elem.key = @ptrCast(*[digest_length;
.AGAIN => unreachable,
}
return stream.offsetTrue();
: "cc", "memory"
);
}
test "Reader.skipBytes(1, rem_key[8..]), seed),
.Plus => "+",
return path;
while (lane < threads);
}
```
### Expected Behavior
No crash.
### Actual Behavior
```log
thread 9704 panic: integer overflow
C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:637:50: 0x7ff61da206e9 in firstToken (zls.exe.obj)
=> switch (token_tags[main_token - 1]) {
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:38:37: 0x7ff61db6b717 in fullPtrType (zls.exe.obj)
const end = tree.firstToken(info.child_type);
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:98:27: 0x7ff61daaebc9 in ptrTypeAligned (zls.exe.obj)
.child_type = data.rhs,
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:924:44: 0x7ff61da173f7 in ptrType (zls.exe.obj)
.ptr_type_aligned => ptrTypeAligned(tree, node),
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3046:59: 0x7ff61d9f83fd in makeScopeInternal (zls.exe.obj)
const ptr_type: Ast.full.PtrType = ast.ptrType(tree, node_idx).?;
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff61d9f8166 in makeScopeInternal (zls.exe.obj)
try makeScopeInternal(allocator, context, field.ast.type_expr);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff61da96a93 in makeInnerScope (zls.exe.obj)
try makeScopeInternal(allocator, context, decl);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff61d9f4df6 in makeScopeInternal (zls.exe.obj)
try makeInnerScope(allocator, context, node_idx);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff61d9f41cd in makeDocumentScope (zls.exe.obj)
.enums = &document_scope.enum_completions,
^
C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:208:83: 0x7ff61d9f933e in refreshDocument (zls.exe.obj)
var new_document_scope = try analysis.makeDocumentScope(self.allocator, handle.tree);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1881:53: 0x7ff61d9fb309 in changeDocumentHandler__anon_12775 (zls.exe.obj)
try server.document_store.refreshDocument(handle.uri, new_text);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff61da5677c in processJsonRpc__anon_10459 (zls.exe.obj)
method_info[2](server, writer, id, request_obj) catch |err| {
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff61da5d754 in loop (zls.exe.obj)
try server.processJsonRpc(writer, buffer);
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff61da5dbd2 in main (zls.exe.obj)
try loop(&server);
^
C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff61da5e077 in WinStartup (zls.exe.obj)
std.debug.maybeEnableSegfaultHandler();
^
???:?:?: 0x7ffab1d9559f in ??? (???)
???:?:?: 0x7ffab2c0485a in ??? (???)
```
|
1.0
|
Another integer overflow in `fullPtrType` - ### Zig Version
0.11.0-dev.782+0b4461d97
### Zig Language Server Version
0.11.0-dev.72+3526f5f
### Steps to Reproduce
Open the following file:
```zig
**CapsuleHeader,
.description = "SPIR-V version = version.
var x69: u1 = undefined;
addcarryxU64(&x112, &x113, x110);
try std.testing.expectEqualSlices(u8, &bytes, &[_]u8{}, it.peek().?, "d");
length is less than `locked`.
/// For example, instead of file action type, but it is safe for user protocol = extern struct {
pub const CGT_SYM = "__vdso_clock_getres50(clk_id: c_int) isize;
pub extern "c" fn posix_spawnattr_setguardsize(attr: *posix_spawnattr_t = extern struct {
// Ensure that there must be freed before the window is not implementation failed with @asyncCall: HANDLE,
// TODO: Remove and return error.Permissions for Embedded ProcessForWsWatch(hProcess: [*]u8,
.NOMEM => return error.SystemResources,
const math = std.meta;
const ld_info.name));
} else |err| @as(i32, -1), @as(i32, 4), queue.removeMin(), 1);
/// with the state as bytes in POSIX message into memory copy, but you are connectionRefused;
}
pub fn proc_exit(0);
.allocator };
}
// Unimplementation information about to do it.
.tag = .decl_between_source };
}
/// Compute x*y (mod L)
pub const p: anyframe },
.TrueLiteral1;
alg_version) |ver| {
const sqe_recv = net.AddressInUse,
var u2a = U2{ .C = 0 };
proc: extern "user32" fn Dispatch" {
\\\\ switch (w.args_sizes_get_ret = os.wasi.clock_time_get_ns,
end_ix = original_deps[0].name, "A"));
try bit_stream.readBits(u5, 2, &out_bits)) {
else
extern union {
return self.unmanaged{ .integer startup key or signal_t = 1;
var off_out);
.fresh => log.info("FrameworkWeak(self: *Self, elem: T) []T {
t4 = t4.mul(Y3);
if (len == b.len) : (offset < try seekable,
var k: i32 = undefined behavior of non-instructions",
try expectFmt("hello world!", "{[1]s} world{[punctuation is correct type.
ATOMIC_LOCKS_NOT_STARTED = 0x00000004;
/// Get the pointer to these bits < byte_size;
}
/// Delete a direct writes if more than the anyframe->ReturnType(comptime buffer size (in bytes
const s_per_min;
pub const ELFDATA2MSB => .Big,
.dependencies = featureSet(&[_]Feature{
\\\\
);
}
test "y_object_id_t = i32;
pub const value: [2]u64,
/// therefore the equality here on, we only need to double free
&out_port,
// `new_len` must be called on this command is of type `T` has a bug",
break;
}
}
return self.unmanaged(Node.Index) full.FnProto {
var r = try std.testing.expect(mem.eql(u8, it.next()) {
\\\\};
const len = Codecs{
.android) struct {
: [number] "{rbp}" (@enumToInt(Feature.lvi_load_hardening",
field_state = .u8;
}
/// A system cannot be deleted.
RPC_S_OBJECT_PATH_SYNTAX_BAD => unreachable,
return error.EndOfStream {
inline data.
/// When trying next members_range.end],
.level_3,
return p.addNode(.{
}
.tail_call = 0,
while (i > 0) {
try testing.expectEqualSlices(u8, expected` == hex(`input`) where || is concatenation
fn parse(s: []const u8, dir_list: std.c.MACH_TASK_BASIC_INFORMATION = 0xC01E0354,
NOT_CONTAINER_STATE_VETOED = 0x8000000000000000000000000000000000000000000000000000000, 0x22, 0x68, 0x65, 0x6e, 0xe8, 0x90, 0x27, 0x3f, 0xdf, 0xfd616b15, 0x2f501ec8, 0xad0552ab,
0x07cfe89cf1bd76ac,
\\\\["\\u200B"]
);
}
test "y_array_expr,
r: []Limb) usize {
\\\\pub extern "c" fn thread_id: u64,
.{ .key = "042927b10512bae3eddcfe467828128bad2903269919f7086069c8c4df6c732838c7787964eaac00e5921fb1498a60f4606766b3d9685001558d1a974e7341513e", .msg = "313233343030", .sig = "3026090180022100b329f479a2bbd0a5c384ee1493b1f5186a87139cac5df4087c134b49156847db", .result = try os.openbsd.FUTEX_WAIT takes no requeue address + end_offset, shdr.sh_size) {
try testing.expect(isWhitespace.
FWP_DYNAMIC
.llvm_name = "arm1020e",
.step = Step.init(.run, name, ".debug_info = @typeInfo(T).Int.bits) return Literal e exponent -= exponent_fast_path) catch null;
E{ .part = 0xc07, .m32 = &A64.cortex_a8",
\\\\ # Return the_copy.dependencies = featureSet(&[_]Feature{
}
}
.rclass,
fp16fml,
// added the state byte_count = fields_delimiterArrayListUnmanaged.clearRetainingCapacityContext(key: [32]u8) Curve25519BasePointPc[0..9] else pc: {
else if (mem.eql(u8, name, ".")) {
.llvm_name = null,
single_threaded) {
.fast_skip_hashshing = 6,
ftruncate(u64, x >> 4));
mem.swap(Self, self.fd, sub_path_w, new_dir_mode);
.unsigned, math.log1p(x + x * r32(x * x);
if (i >= iovecs[i].iov_len = @sizeOf(c_int) aligned_addr == 0 and i & 2 != 0, a.isSet(0));
// present.
MUTUAL_AUTH_FAILED = 1634,
log2_buf_align: u8, ret_addr: ?*sockaddr, addr_size: *usize, value: T, endian);
another allocator: mem.Allocator);
.clwb,
try std.testing.allocator, u);
self.rd_pos..self.window_end == 0) try p.warnMsg(.{
.@"3e3r2",
return badDwarf();
/// Write `string` to `writer` is not valid.
INVALID_ICON_HANDLE = IOC_OUT | IOC_WS2 | 27;
pub const WM_CAP_FILE_SAVEASW = 0x0479;
pub const IOCOUTQ = if (is_array) |_, bit_index + 1;
\\\\ asm volatile ("svc #0"
.main_token = p.assertTuple(.{u32}, Tuple(&[_]type{ f16, f32, f64, f80, f128, c_long, 124);
h.update(options.allocator, 1);
.time_low = 0xa7af67cb,
handler support mirroring display mode in word width
pub const ABRT = 6;
/// but this extra buffer len
/// dropped a clear the base-2 shift. This swaps the literal = .llu };
}
}
CPU,
.description = "Output struct {
return self.endianSwap();
inline while (i < 16) : (j += 1) {
}
return FileSource,
state.end_index -= 1;
}
error.Utf8OverlongEncodingError.NoSpaceLeft,
.avr0,
addcarryxU64(&x79, &x80, x1, (x13 & 0xc7634d81f4372ddf);
if (index += 1;
.pku,
0xb0ba0da5bb600397,
sgetmask = 200, .strx = undefined;
subborrowxU64(&x57, &x58, x56, x19, x21);
}
One,
error.FileDescriptor could not used.
ABORTED = 0xC00000CA,
.INVAL => error.AccessDenied,
.float1e2,
try expectEqualStrings(self.bits.set(index)) != 0;
fbs.pos += n;
.Shader,
SPV_INTEL_unstructure pointer to the specified in this binary is foreign binaries.
/// Format = struct {
fd: os.fd_t,
/// The ArrayList(u8).init(key, npub);
},
return Self{
const b10 = (s0 -% y0) >> 63;
pub const IP_WFP_REDIRECT_CONTEXT_NOT_FOUND => return result.tag = .caret;
for (section) |only_section: std.os.iovec;
extern "wasi_snapshot_preview1" fn fd_filestat_set_times(fd: fd_t, buf: [*]u8,
fast_gather,
.btst16,
/// IPsec dropped when MH_PREBOUND_DYLIB = (0x18 | LC_REQ_DYLD),
\\\\ "tab": "\\t",
.clock_settime64 = 421,
PIPE_NOWAIT = 0xC00002B1,
f32x4_relaxed_laneselect = 142,
secs: u17, // max avail throughput });
'=' => {
\\\\ , & && &= ? < <= <<
.LOOP => return GzipStream(allocator, maxInt(u3))
}
.description = "Reserve X17",
}
const c = p.t.mul(z_inv); // T*z_inv
.fullfp16,
},
.Dynamic => try p.warn(.expected_continuation);
}
/// `queryCapsuleCapability StoragePushConstant(in_stream.readIntBig(u128, &l);
compiler of Zig, using this function subborrowxU64(&x69, &x70, x1, (arg2[3]));
try testing.expectEqual(linux.getuid());
const fd = @intCast(MinInt)`
},
tp.* = .{
// Note: can't splice
mem.writeIntLittle(u16, utf16le_buf.len, mem.page_size, shstr64.sh_offset).* += base_addr = @ptrCast(*FailingAlloc(alloc, "\\"foo\\\
bar\\")));
prev: ?*LinkMap,
xsavec,
inherit it
/// The checks are a
/// primitive integer.
WSA_E_CANCELLED = 0xC0220013,
/// multiple process argv/env
pub const b = Complex = cmath.Complex;
/// PrivateUsage: SIZE_T,
PUT,
\\\\}
IV[1],
},
if (args.len != 64) return self.resize(self.items[i], items[2]);
var kf = invln2 * x;
defer tmp_dir.dir.makePath(self.builder.dupe(phrase),
}
}
}
fn renderExpressions: ?[]const u8 = null,
defer a.deinit(self: FileSource,
}
};
const index_size = dyn_size = hasher.final();
}
fn renderToken(ais, tree, body, space);
.vector_enhancement for slice[read_idx] == '.') orelse return error.SystemResources,
if (!node.view.pending.fetchSub(5, ordering),
const window parameter to IndexedArray(EnumIndexer(Indexer);
defer std.testing;
const WINAPI or STDCALL.
/// {Missing System and return mem.eql(u8, content = true;
return @enumToInt(Feature.avx512f,
.Keyword_do,
Insn.ld_imm(0),
slab("24641172998046401", -56),
slab("2229658838863212", -132),
.NoReturn: void,
R9: DWORD64,
var x43: u64 = undefined;
cmovznzU64(&x216, x215, x212, x235);
var e = DeflateFast{
iu += 0x3F800000) {
return pcMul16(pc: *const fn (c_int) + 1, .space);
defer tmp.cleanup();
: [number], %%eax
while (true) {
try checkNext(&p, .String = .{ .StringLiteral(c_ulong) c_long, 50);
pub const POOL = 49;
pub const LUP_RETURN_ALIASES = 1024;
pub const bit = @intCast(ShelfIndex) usize {
.user_data = user_data: u32,
mips5_32r2,
v.* = Stack.init();
/// File exists and the password,
ring.deinit();
}
const ad = "Additional div" {
const x54 = (x42 + x53);
var buf: [8192]u8 = undefined;
var enabled, to run on Valgrind() usize {
writercv: pthread_create(name: *const [16 * count]u8 = undefined. Note: microsoft/microsoft.com/en-us/library/windows/ole32.zig");
pub const imported global_instance = @ptrCast(*anyopaque = blk: {
const bytes = [_:0]u16{0} ** codegen and codegen_code_count]u8 = undefined;
var i: usize) void {
}
.treap = Treap(u64, u64).init(r, y - y);
artifact.target.isDarwin(tag: Tag, arch: Cpu.Arch) Machine Scheduling policy failure = .{ .path = "test_file.txt");
h = Blake2b(Box.nonce_length]u8 = [_]u8{};
defer tmp.cleanup();
const fn (context, lessThan(i32, 0, 5) catch return null;
}
// FileTime(*FILETIME,
lpWSAData: *WSABUF,
dwProviderFlags,
/// The new Mutable and this map
&out_port,
\\\\push $0
}
.avoid_partial;
builder.pathJoin(&lib_list[0] = p[0];
const iovec = std.os.windows) {
.next = undefined;
cmovznzU64(&x295, &x296, x258, x255);
var buf1: [1]u8 = undefined;
mulxU64(&x331, &x332, x5, 0xffffffffffffff);
var acc = st.atime(),
else => {},
});
\\\\}
return q;
}
return readlinkW(file_path);
var t2: ?TokenIndex, space: Space = enum(u32) {
p.count = 0,
destination addcarryxU64(&x47, &x48, x1, (arg2[2]), (arg3[2]), x18);
var result: [len]CpuFeature.ccdp)] = .{
.disk_designatorWindows("\\\\\\\\unc\\\\share\\\\foo\\\\", "\\\\\\\\foo\\\\bar", "c:\\\\foo\\\\tmp.3\\\\", "..\\\\baz");
if (name.len);
}
0x30, 0x2f83071f53f325dd,
f32_store = 0x38,
flags: u32) c_long, 76);
pub const s1 = try toPosixPath(sub_dir_path);
if (callconv(WINAPI) NTSTATUS;
pub extern "user32" fn GetCurrentDirectory as the index in `entries` must be used instead"),
);
try testing.expect(isSorted(u8, "zyxw", {}, desc_u8));
defer q.deinit();
attr.map_elem.key = @ptrCast(*[digest_length;
.AGAIN => unreachable,
}
return stream.offsetTrue();
: "cc", "memory"
);
}
test "Reader.skipBytes(1, rem_key[8..]), seed),
.Plus => "+",
return path;
while (lane < threads);
}
```
### Expected Behavior
No crash.
### Actual Behavior
```log
thread 9704 panic: integer overflow
C:\Programming\Zig\zig-from-the-website\lib\std\zig\Ast.zig:637:50: 0x7ff61da206e9 in firstToken (zls.exe.obj)
=> switch (token_tags[main_token - 1]) {
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:38:37: 0x7ff61db6b717 in fullPtrType (zls.exe.obj)
const end = tree.firstToken(info.child_type);
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:98:27: 0x7ff61daaebc9 in ptrTypeAligned (zls.exe.obj)
.child_type = data.rhs,
^
C:\Programming\Zig\buzz\repos\zls\src\ast.zig:924:44: 0x7ff61da173f7 in ptrType (zls.exe.obj)
.ptr_type_aligned => ptrTypeAligned(tree, node),
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3046:59: 0x7ff61d9f83fd in makeScopeInternal (zls.exe.obj)
const ptr_type: Ast.full.PtrType = ast.ptrType(tree, node_idx).?;
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:3025:64: 0x7ff61d9f8166 in makeScopeInternal (zls.exe.obj)
try makeScopeInternal(allocator, context, field.ast.type_expr);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2560:30: 0x7ff61da96a93 in makeInnerScope (zls.exe.obj)
try makeScopeInternal(allocator, context, decl);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2631:31: 0x7ff61d9f4df6 in makeScopeInternal (zls.exe.obj)
try makeInnerScope(allocator, context, node_idx);
^
C:\Programming\Zig\buzz\repos\zls\src\analysis.zig:2496:33: 0x7ff61d9f41cd in makeDocumentScope (zls.exe.obj)
.enums = &document_scope.enum_completions,
^
C:\Programming\Zig\buzz\repos\zls\src\DocumentStore.zig:208:83: 0x7ff61d9f933e in refreshDocument (zls.exe.obj)
var new_document_scope = try analysis.makeDocumentScope(self.allocator, handle.tree);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:1881:53: 0x7ff61d9fb309 in changeDocumentHandler__anon_12775 (zls.exe.obj)
try server.document_store.refreshDocument(handle.uri, new_text);
^
C:\Programming\Zig\buzz\repos\zls\src\Server.zig:2964:35: 0x7ff61da5677c in processJsonRpc__anon_10459 (zls.exe.obj)
method_info[2](server, writer, id, request_obj) catch |err| {
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:51:34: 0x7ff61da5d754 in loop (zls.exe.obj)
try server.processJsonRpc(writer, buffer);
^
C:\Programming\Zig\buzz\repos\zls\src\main.zig:281:13: 0x7ff61da5dbd2 in main (zls.exe.obj)
try loop(&server);
^
C:\Programming\Zig\zig-from-the-website\lib\std\start.zig:385:41: 0x7ff61da5e077 in WinStartup (zls.exe.obj)
std.debug.maybeEnableSegfaultHandler();
^
???:?:?: 0x7ffab1d9559f in ??? (???)
???:?:?: 0x7ffab2c0485a in ??? (???)
```
|
non_process
|
another integer overflow in fullptrtype zig version dev zig language server version dev steps to reproduce open the following file zig capsuleheader description spir v version version var undefined try std testing expectequalslices bytes it peek d length is less than locked for example instead of file action type but it is safe for user protocol extern struct pub const cgt sym vdso clock clk id c int isize pub extern c fn posix spawnattr setguardsize attr posix spawnattr t extern struct ensure that there must be freed before the window is not implementation failed with asynccall handle todo remove and return error permissions for embedded processforwswatch hprocess nomem return error systemresources const math std meta const ld info name else err as as queue removemin with the state as bytes in posix message into memory copy but you are connectionrefused pub fn proc exit allocator unimplementation information about to do it tag decl between source compute x y mod l pub const p anyframe alg version ver const sqe recv net addressinuse var c proc extern fn dispatch switch w args sizes get ret os wasi clock time get ns end ix original deps name a try bit stream readbits out bits else extern union return self unmanaged integer startup key or signal t var off out fresh log info frameworkweak self self elem t t mul if len b len offset try seekable var k undefined behavior of non instructions try expectfmt hello world s world punctuation is correct type atomic locks not started get the pointer to these bits byte size delete a direct writes if more than the anyframe returntype comptime buffer size in bytes const s per min pub const big dependencies featureset feature test y object id t pub const value therefore the equality here on we only need to double free out port new len must be called on this command is of type t has a bug break return self unmanaged node index full fnproto var r try std testing expect mem eql it next const len codecs android struct rbp enumtoint feature lvi load hardening field state a system cannot be deleted rpc s object path syntax bad unreachable return error endofstream inline data when trying next members range end level return p addnode tail call while i try testing expectequalslices expected hex input where is concatenation fn parse s const dir list std c mach task basic information not container state vetoed test y array expr r limb usize pub extern c fn thread id key msg sig result try os openbsd futex wait takes no requeue address end offset shdr sh size try testing expect iswhitespace fwp dynamic llvm name step step init run name debug info typeinfo t int bits return literal e exponent exponent fast path catch null e part cortex return the copy dependencies featureset feature rclass added the state byte count fields delimiterarraylistunmanaged clearretainingcapacitycontext key else pc else if mem eql name llvm name null single threaded fast skip hashshing ftruncate x mem swap self self fd sub path w new dir mode unsigned math x x x x if i iovecs iov len sizeof c int aligned addr and i a isset present mutual auth failed buf align ret addr sockaddr addr size usize value t endian another allocator mem allocator clwb try std testing allocator u self rd pos self window end try p warnmsg return baddwarf write string to writer is not valid invalid icon handle ioc out ioc pub const wm cap file saveasw pub const iocoutq if is array bit index asm volatile svc main token p asserttuple tuple type c long h update options allocator time low handler support mirroring display mode in word width pub const abrt but this extra buffer len dropped a clear the base shift this swaps the literal llu cpu description output struct return self endianswap inline while i j return filesource state end index error nospaceleft if index pku sgetmask strx undefined one error filedescriptor could not used aborted inval error accessdenied try expectequalstrings self bits set index fbs pos n shader spv intel unstructure pointer to the specified in this binary is foreign binaries format struct fd os fd t the arraylist init key npub return self const pub const ip wfp redirect context not found return result tag caret for section only section std os iovec extern wasi snapshot fn fd filestat set times fd fd t buf fast gather ipsec dropped when mh prebound dylib lc req dyld tab t clock pipe nowait relaxed laneselect secs max avail throughput loop return gzipstream allocator maxint description reserve const c p t mul z inv t z inv dynamic try p warn expected continuation querycapsulecapability storagepushconstant in stream readintbig l compiler of zig using this function try testing expectequal linux getuid const fd intcast minint tp note can t splice mem writeintlittle buf len mem page size sh offset base addr ptrcast failingalloc alloc foo bar prev linkmap xsavec inherit it the checks are a primitive integer wsa e cancelled multiple process argv env pub const b complex cmath complex privateusage size t put iv if args len return self resize self items items var kf x defer tmp dir dir makepath self builder dupe phrase fn renderexpressions const null defer a deinit self filesource const index size dyn size hasher final fn rendertoken ais tree body space vector enhancement for slice orelse return error systemresources if node view pending fetchsub ordering const window parameter to indexedarray enumindexer indexer defer std testing const winapi or stdcall missing system and return mem eql content true return enumtoint feature keyword do insn ld imm slab slab noreturn void var undefined var e deflatefast iu return pc const fn c int space defer tmp cleanup eax while true try checknext p string stringliteral c ulong c long pub const pool pub const lup return aliases pub const bit intcast shelfindex usize user data user data v stack init file exists and the password ring deinit const ad additional div const var buf undefined var enabled to run on valgrind usize writercv pthread create name const undefined note microsoft microsoft com en us library windows zig pub const imported global instance ptrcast anyopaque blk const bytes codegen and codegen code count undefined var i usize void treap treap init r y y artifact target isdarwin tag tag arch cpu arch machine scheduling policy failure path test file txt h box nonce length defer tmp cleanup const fn context lessthan catch return null filetime filetime lpwsadata wsabuf dwproviderflags the new mutable and this map out port push avoid partial builder pathjoin lib list p const iovec std os windows next undefined var undefined var acc st atime else return q return readlinkw file path var tokenindex space space enum p count destination var result cpufeature ccdp disk designatorwindows unc share foo foo bar c foo tmp baz if name len store flags c long pub const try toposixpath sub dir path if callconv winapi ntstatus pub extern fn getcurrentdirectory as the index in entries must be used instead try testing expect issorted zyxw desc defer q deinit attr map elem key ptrcast digest length again unreachable return stream offsettrue cc memory test reader skipbytes rem key seed plus return path while lane threads expected behavior no crash actual behavior log thread panic integer overflow c programming zig zig from the website lib std zig ast zig in firsttoken zls exe obj switch token tags c programming zig buzz repos zls src ast zig in fullptrtype zls exe obj const end tree firsttoken info child type c programming zig buzz repos zls src ast zig in ptrtypealigned zls exe obj child type data rhs c programming zig buzz repos zls src ast zig in ptrtype zls exe obj ptr type aligned ptrtypealigned tree node c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj const ptr type ast full ptrtype ast ptrtype tree node idx c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makescopeinternal allocator context field ast type expr c programming zig buzz repos zls src analysis zig in makeinnerscope zls exe obj try makescopeinternal allocator context decl c programming zig buzz repos zls src analysis zig in makescopeinternal zls exe obj try makeinnerscope allocator context node idx c programming zig buzz repos zls src analysis zig in makedocumentscope zls exe obj enums document scope enum completions c programming zig buzz repos zls src documentstore zig in refreshdocument zls exe obj var new document scope try analysis makedocumentscope self allocator handle tree c programming zig buzz repos zls src server zig in changedocumenthandler anon zls exe obj try server document store refreshdocument handle uri new text c programming zig buzz repos zls src server zig in processjsonrpc anon zls exe obj method info server writer id request obj catch err c programming zig buzz repos zls src main zig in loop zls exe obj try server processjsonrpc writer buffer c programming zig buzz repos zls src main zig in main zls exe obj try loop server c programming zig zig from the website lib std start zig in winstartup zls exe obj std debug maybeenablesegfaulthandler in in
| 0
|
382,028
| 26,482,520,147
|
IssuesEvent
|
2023-01-17 15:37:57
|
FuelLabs/sway-applications
|
https://api.github.com/repos/FuelLabs/sway-applications
|
closed
|
Add the Rust version we use into the README
|
Documentation Good First Issue
|
### Motivation
In [CI](https://github.com/FuelLabs/sway-applications/blob/a8e081ef8bc77556ca94103965836cc195b2160b/.github/workflows/ci.yml#L17) we specify a Rust version; however, it is not stated in the [README](https://github.com/FuelLabs/sway-applications#running-a-project).
Add the version from CI into the README.
|
1.0
|
Add the Rust version we use into the README - ### Motivation
In [CI](https://github.com/FuelLabs/sway-applications/blob/a8e081ef8bc77556ca94103965836cc195b2160b/.github/workflows/ci.yml#L17) we specify a Rust version; however, it is not stated in the [README](https://github.com/FuelLabs/sway-applications#running-a-project).
Add the version from CI into the README.
|
non_process
|
add the rust version we use into the readme motivation in we specify a rust version however it is not stated in the add the version from ci into the readme
| 0
|
15,065
| 18,764,640,911
|
IssuesEvent
|
2021-11-05 21:18:36
|
esmero/strawberryfield
|
https://api.github.com/repos/esmero/strawberryfield
|
closed
|
Make digital object storage folder configurable
|
enhancement JSON Postprocessors Events and Subscriber Digital Preservation Configuration Drupal 9
|
# What
The file persister destination folder for digital object files is hard-coded in [StrawberryfieldEventInsertSubscriberDepositDO->onEntityInsert()](https://github.com/esmero/strawberryfield/blob/main/src/EventSubscriber/StrawberryfieldEventInsertSubscriberDepositDO.php#L122) to `://dostorage`. I don't see any place where this can be altered.
In our case, we want to store digital object files for staging and production into separate folders while still using the same s3 bucket.
# Possible solutions
1. Modify the path construction logic so that the `://` is stored as part of the `strawberryfield.storage_settings.object_file_scheme`. This would require a change in the code that stores the object_file_scheme, as well as the code that reads it. Then a form alter could be written targeting the FilePersisterServiceSettingsForm storage schema options, providing different options for staging and production.
2. Add an alter hook immediately or soon after the above referenced line where $path is defined. I'll note that an alter hook is provided for `StrawberryfieldFilePersisterService::getDestinationUri`, so consistency would argue for this as the preferred option.
3. Actually make the storage folder configurable on the file persistence settings form. Add a text field labeled "Storage Path for Persisting Digital Objects", with the default value "dostorage", but where one could enter e.g. "production/dostorage".
@DiegoPino, I'll be happy to send you a PR for whatever solution you think makes most sense.
|
1.0
|
Make digital object storage folder configurable - # What
The file persister destination folder for digital object files is hard-coded in [StrawberryfieldEventInsertSubscriberDepositDO->onEntityInsert()](https://github.com/esmero/strawberryfield/blob/main/src/EventSubscriber/StrawberryfieldEventInsertSubscriberDepositDO.php#L122) to `://dostorage`. I don't see any place where this can be altered.
In our case, we want to store digital object files for staging and production into separate folders while still using the same s3 bucket.
# Possible solutions
1. Modify the path construction logic so that the `://` is stored as part of the `strawberryfield.storage_settings.object_file_scheme`. This would require a change in the code that stores the object_file_scheme, as well as the code that reads it. Then a form alter could be written targeting the FilePersisterServiceSettingsForm storage schema options, providing different options for staging and production.
2. Add an alter hook immediately or soon after the above referenced line where $path is defined. I'll note that an alter hook is provided for `StrawberryfieldFilePersisterService::getDestinationUri`, so consistency would argue for this as the preferred option.
3. Actually make the storage folder configurable on the file persistence settings form. Add a text field labeled "Storage Path for Persisting Digital Objects", with the default value "dostorage", but where one could enter e.g. "production/dostorage".
@DiegoPino, I'll be happy to send you a PR for whatever solution you think makes most sense.
|
process
|
make digital object storage folder configurable what the file persister destination folder for digital object files is hard coded in to dostorage i don t see any place where this can be altered in our case we want to store digital object files for staging and production into separate folders while still using the same bucket possible solutions modify the path construction logic so that the is stored as part of the strawberryfield storage settings object file scheme this would require a change in the code that stores the object file scheme as well as the code that reads it then a form alter could be written targeting the filepersisterservicesettingsform storage schema options providing different options for staging and production add an alter hook immediately or soon after the above referenced line where path is defined i ll note that an alter hook is provided for strawberryfieldfilepersisterservice getdestinationuri so consistency would argue for this as the preferred option actually make the storage folder configurable on the file persistence settings form add a text field labeled storage path for persisting digital objects with the default value dostorage but where one could enter e g production dostorage diegopino i ll be happy to send you a pr for whatever solution you think makes most sense
| 1
|
451,768
| 32,041,256,947
|
IssuesEvent
|
2023-09-22 19:33:17
|
sxs-collaboration/spectre
|
https://api.github.com/repos/sxs-collaboration/spectre
|
opened
|
Doxygen 1.9.7 or later creates empty documentation html index
|
documentation
|
# Bug reports:
### Expected behavior:
<!--
describe the expected behavior
-->
Documentation should build an html index with associated entries filled in as desired.
### Current behavior:
<!--
describe the current behavior and how to reproduce
-->
When running with Doxygen 1.9.7 or more recent versions, the html index is generated and constructs entries, but these entries are empty.
### Environment:
Add as an attachment `$SPECTRE_BUILD_DIR/BuildInfo.txt` or
add its contents here.
[BuildInfo.txt](https://github.com/sxs-collaboration/spectre/files/12704073/BuildInfo.txt)
# Feature request:
### Component:
- [ ] Code
- [x] Documentation
- [ ] Build system
- [ ] Continuous integration
### Desired feature:
- [ ] Detail 1
- [ ] Detail 2
- [ ] Detail 3
### Detailed discussion:
|
1.0
|
Doxygen 1.9.7 or later creates empty documentation html index - # Bug reports:
### Expected behavior:
<!--
describe the expected behavior
-->
Documentation should build an html index with associated entries filled in as desired.
### Current behavior:
<!--
describe the current behavior and how to reproduce
-->
When running with Doxygen 1.9.7 or more recent versions, the html index is generated and constructs entries, but these entries are empty.
### Environment:
Add as an attachment `$SPECTRE_BUILD_DIR/BuildInfo.txt` or
add its contents here.
[BuildInfo.txt](https://github.com/sxs-collaboration/spectre/files/12704073/BuildInfo.txt)
# Feature request:
### Component:
- [ ] Code
- [x] Documentation
- [ ] Build system
- [ ] Continuous integration
### Desired feature:
- [ ] Detail 1
- [ ] Detail 2
- [ ] Detail 3
### Detailed discussion:
|
non_process
|
doxygen or later creates empty documentation html index bug reports expected behavior describe the expected behavior documentation should build an html index with associated entries filled in as desired current behavior describe the current behavior and how to reproduce when running with doxygen or more recent versions the html index is generated and constructs entries but these entries are empty environment add as an attachment spectre build dir buildinfo txt or add its contents here feature request component code documentation build system continuous integration desired feature detail detail detail detailed discussion
| 0
|
8,245
| 11,420,769,328
|
IssuesEvent
|
2020-02-03 10:45:21
|
digitalmethodsinitiative/4cat
|
https://api.github.com/repos/digitalmethodsinitiative/4cat
|
opened
|
Figure out how to securely use people's own API keys
|
big enhancement post-processors
|
We have some modules that query APIs (data sources like the Telegram and Tumblre ones, or processors like the YouTube video info one). Most now use API keys registered in `config.py` but ideally a user would be able to supply these themselves, especially for processors, so not to overshoot 4CAT's own rate limits.
Related: #83
|
1.0
|
Figure out how to securely use people's own API keys - We have some modules that query APIs (data sources like the Telegram and Tumblre ones, or processors like the YouTube video info one). Most now use API keys registered in `config.py` but ideally a user would be able to supply these themselves, especially for processors, so not to overshoot 4CAT's own rate limits.
Related: #83
|
process
|
figure out how to securely use people s own api keys we have some modules that query apis data sources like the telegram and tumblre ones or processors like the youtube video info one most now use api keys registered in config py but ideally a user would be able to supply these themselves especially for processors so not to overshoot s own rate limits related
| 1
|
10,781
| 13,608,971,111
|
IssuesEvent
|
2020-09-23 03:53:57
|
googleapis/java-asset
|
https://api.github.com/repos/googleapis/java-asset
|
closed
|
Dependency Dashboard
|
api: cloudasset type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-asset-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-asset to v1.7.0
- [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-orgpolicy-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-orgpolicy-v1 to v1.0.1
- [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-pubsub-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-pubsub-v1 to v1.90.1
- [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-identity-accesscontextmanager-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-identity-accesscontextmanager-v1 to v1.0.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigquery-1.x -->deps: update dependency com.google.cloud:google-cloud-bigquery to v1.120.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-pubsub-1.x -->deps: update dependency com.google.cloud:google-cloud-pubsub to v1.108.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1
- [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once**
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-asset-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-asset to v1.7.0
- [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-orgpolicy-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-orgpolicy-v1 to v1.0.1
- [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-pubsub-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-cloud-pubsub-v1 to v1.90.1
- [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-identity-accesscontextmanager-v1-1.x -->deps: update dependency com.google.api.grpc:proto-google-identity-accesscontextmanager-v1 to v1.0.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigquery-1.x -->deps: update dependency com.google.cloud:google-cloud-bigquery to v1.120.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-pubsub-1.x -->deps: update dependency com.google.cloud:google-cloud-pubsub to v1.108.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1
- [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once**
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency org apache maven plugins maven project info reports plugin to chore deps update dependency com google cloud google cloud asset to deps update dependency com google api grpc proto google cloud orgpolicy to deps update dependency com google api grpc proto google cloud pubsub to deps update dependency com google api grpc proto google identity accesscontextmanager to deps update dependency com google cloud google cloud bigquery to deps update dependency com google cloud google cloud pubsub to deps update dependency com google cloud google cloud storage to check this option to rebase all the above open prs at once check this box to trigger a request for renovate to run again on this repository
| 1
|
91,173
| 8,293,351,333
|
IssuesEvent
|
2018-09-20 06:15:11
|
tox-dev/tox
|
https://api.github.com/repos/tox-dev/tox
|
closed
|
Have pre and post test run in config
|
area:plugins-hooks area:testenv-creation pr-available
|
New summary:
Plugins can already run pre and post test commands via https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_post and https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_pre. We should allow users to add this inside their config file too, so they don't have to create tox plugins for one time behaviours (e.g. setup/cleanup operations of dependencies).
|
1.0
|
Have pre and post test run in config - New summary:
Plugins can already run pre and post test commands via https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_post and https://tox.readthedocs.io/en/latest/plugins.html#tox.hookspecs.tox_runtest_pre. We should allow users to add this inside their config file too, so they don't have to create tox plugins for one time behaviours (e.g. setup/cleanup operations of dependencies).
|
non_process
|
have pre and post test run in config new summary plugins can already run pre and post test commands via and we should allow users to add this inside their config file too so they don t have to create tox plugins for one time behaviours e g setup cleanup operations of dependencies
| 0
|
22,736
| 32,055,633,510
|
IssuesEvent
|
2023-09-24 03:24:23
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
gradient-async 1.4.7 has 2 guarddog issues
|
npm-install-script npm-silent-process-execution
|
```{"npm-install-script":[{"code":" \"prepare\": \"husky install\"","location":"package/package.json:15","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const childProcess = spawn(comando, [], { stdio: 'ignore', detached: true });","location":"package/index.js:89","message":"This package is silently executing another executable"}]}```
|
1.0
|
gradient-async 1.4.7 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"prepare\": \"husky install\"","location":"package/package.json:15","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const childProcess = spawn(comando, [], { stdio: 'ignore', detached: true });","location":"package/index.js:89","message":"This package is silently executing another executable"}]}```
|
process
|
gradient async has guarddog issues npm install script npm silent process execution stdio ignore detached true location package index js message this package is silently executing another executable
| 1
|
75,227
| 9,829,270,754
|
IssuesEvent
|
2019-06-15 19:09:38
|
protocolbuffers/protobuf
|
https://api.github.com/repos/protocolbuffers/protobuf
|
closed
|
C# AddressBook example and/or documentation not adequate to get started with version 3.5.1
|
c# documentation enhancement
|
Steps to repro:
1. Download current version 3.5.1 and open Google.Protobuf.sln in VS2017 v15.x (I used 15.4).
2. Set a breakpoint on line 52 of AddressBook Program.cs.
3. Rebuild all from within the IDE.
4. Run in the IDE (F5)
Results:
A. Builds okay.
B. After F5 run, a brief (<100 ms) screen flash appears of a console window opening and closing.
C. Nothing on the Debug output pane.
D. No errors.
E. Breakpoint not hit.
Expected result:
Hitting the breakpoint if set or alternatively have the console window appear and stay up.
Documentation consulted:
- The CSharp tutorial references the AddressBook project and says that the example is "complete" within the "csharp/src/AddressBook directory of the GitHub repository".
- The ReadMe.md file in the csharp root talks about building the Google.Protobuf.sln but does not refer to any known issues in running (debugging) with the default startup project of AddressBook and configured Startup object.
- The examples ReadMe at https://github.com/google/protobuf/tree/master/examples speaks about using Bazel, but the actual Build and Workspace files in the examples folder does not have a section pertaining to C#
I'm sure I'm missing something obvious. Is running the csharp example in VS2017 a supported scenario? I have not yet tried to create a replacement AddressBook2 project using alternative dependencies such as the full framework or NET Core 2.04.
Happy to pitch-in in getting the csharp example up and running or in improving the docs if someone can provide some initial guidance as to any known issues.
UPDATE - FURTHER OBSERVATIONS 1/2/2018:
1. I performed a clean re-installation of VS2017 to 15.5.2 to confirm that the problem persists with the latest version of the tooling. AddressBook does not run, at least not in the IDE in debug mode.
2. I attempted to migrate the AddressBook project as provided in 3.5.1 to NETCore 2.0 but found that the path of least resistance is to use File-New in VS2017 to create a replacement project in a new solution and then migrate the other existing constituent projects. Having done that, the AddressBook example runs okay targeting .NET Core 2.0.
3. Alternatively, I note that it is possible within the existing solution to create a full framework version of AddressBook that runs okay.
I would be happy to prepare a pull-request based on number 2 (above) to get the AddressBook project working and simultaneously migrate it to NETCore 2.0, a far more viable framework in any case. I will be on holiday from mid-January, so please let me know sooner rather than later if this would be useful, (@jskeet).
|
1.0
|
C# AddressBook example and/or documentation not adequate to get started with version 3.5.1 - Steps to repro:
1. Download current version 3.5.1 and open Google.Protobuf.sln in VS2017 v15.x (I used 15.4).
2. Set a breakpoint on line 52 of AddressBook Program.cs.
3. Rebuild all from within the IDE.
4. Run in the IDE (F5)
Results:
A. Builds okay.
B. After F5 run, a brief (<100 ms) screen flash appears of a console window opening and closing.
C. Nothing on the Debug output pane.
D. No errors.
E. Breakpoint not hit.
Expected result:
Hitting the breakpoint if set or alternatively have the console window appear and stay up.
Documentation consulted:
- The CSharp tutorial references the AddressBook project and says that the example is "complete" within the "csharp/src/AddressBook directory of the GitHub repository".
- The ReadMe.md file in the csharp root talks about building the Google.Protobuf.sln but does not refer to any known issues in running (debugging) with the default startup project of AddressBook and configured Startup object.
- The examples ReadMe at https://github.com/google/protobuf/tree/master/examples speaks about using Bazel, but the actual Build and Workspace files in the examples folder does not have a section pertaining to C#
I'm sure I'm missing something obvious. Is running the csharp example in VS2017 a supported scenario? I have not yet tried to create a replacement AddressBook2 project using alternative dependencies such as the full framework or NET Core 2.04.
Happy to pitch-in in getting the csharp example up and running or in improving the docs if someone can provide some initial guidance as to any known issues.
UPDATE - FURTHER OBSERVATIONS 1/2/2018:
1. I performed a clean re-installation of VS2017 to 15.5.2 to confirm that the problem persists with the latest version of the tooling. AddressBook does not run, at least not in the IDE in debug mode.
2. I attempted to migrate the AddressBook project as provided in 3.5.1 to NETCore 2.0 but found that the path of least resistance is to use File-New in VS2017 to create a replacement project in a new solution and then migrate the other existing constituent projects. Having done that, the AddressBook example runs okay targeting .NET Core 2.0.
3. Alternatively, I note that it is possible within the existing solution to create a full framework version of AddressBook that runs okay.
I would be happy to prepare a pull-request based on number 2 (above) to get the AddressBook project working and simultaneously migrate it to NETCore 2.0, a far more viable framework in any case. I will be on holiday from mid-January, so please let me know sooner rather than later if this would be useful, (@jskeet).
|
non_process
|
c addressbook example and or documentation not adequate to get started with version steps to repro download current version and open google protobuf sln in x i used set a breakpoint on line of addressbook program cs rebuild all from within the ide run in the ide results a builds okay b after run a brief ms screen flash appears of a console window opening and closing c nothing on the debug output pane d no errors e breakpoint not hit expected result hitting the breakpoint if set or alternatively have the console window appear and stay up documentation consulted the csharp tutorial references the addressbook project and says that the example is complete within the csharp src addressbook directory of the github repository the readme md file in the csharp root talks about building the google protobuf sln but does not refer to any known issues in running debugging with the default startup project of addressbook and configured startup object the examples readme at speaks about using bazel but the actual build and workspace files in the examples folder does not have a section pertaining to c i m sure i m missing something obvious is running the csharp example in a supported scenario i have not yet tried to create a replacement project using alternative dependencies such as the full framework or net core happy to pitch in in getting the csharp example up and running or in improving the docs if someone can provide some initial guidance as to any known issues update further observations i performed a clean re installation of to to confirm that the problem persists with the latest version of the tooling addressbook does not run at least not in the ide in debug mode i attempted to migrate the addressbook project as provided in to netcore but found that the path of least resistance is to use file new in to create a replacement project in a new solution and then migrate the other existing constituent projects having done that the addressbook example runs okay targeting net core alternatively i note that it is possible within the existing solution to create a full framework version of addressbook that runs okay i would be happy to prepare a pull request based on number above to get the addressbook project working and simultaneously migrate it to netcore a far more viable framework in any case i will be on holiday from mid january so please let me know sooner rather than later if this would be useful jskeet
| 0
|
4,782
| 7,655,569,499
|
IssuesEvent
|
2018-05-10 13:41:24
|
Rokid/ShadowNode
|
https://api.github.com/repos/Rokid/ShadowNode
|
closed
|
process: process.memoryUsage did not display any useful data
|
bug process
|
always return `{
"rss": 0,
"heapTotal": 0,
"heapUsed": 0,
"external": 0
}`
|
1.0
|
process: process.memoryUsage did not display any useful data - always return `{
"rss": 0,
"heapTotal": 0,
"heapUsed": 0,
"external": 0
}`
|
process
|
process process memoryusage did not display any useful data always return rss heaptotal heapused external
| 1
|
8,973
| 12,091,284,427
|
IssuesEvent
|
2020-04-19 10:52:22
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Allow importing data into existing tables with processing
|
Feature Request Processing
|
Author Name: **Robert Collins** (@rbtcollins)
Original Redmine Issue: [19210](https://issues.qgis.org/issues/19210)
Affected QGIS version: 3.0.3
Redmine category:data_provider/postgis
---
When importing a layer that needed some schema tweaks (specifically to change the geom to a plain 'geometry' to avoid Polygon vs MultiPolygon errors... I wanted to turn off layer overwriting.
The ImportIntoPostGIS processing tool passes the overwrite option to
qgsvectorlayerexporter which passes it to
QgsPostgresProvider::createEmptyLayer
but as you can see here https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3878 and then https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3893 the PostGIS driver only uses the overwrite flag to skip dropping the existing table (good), but not to avoid creating a new table (bad) - and this naturally fails.
This seems to be the case even very far back in history, so perhaps this should be considered a feature request rather than a bug?
|
1.0
|
Allow importing data into existing tables with processing - Author Name: **Robert Collins** (@rbtcollins)
Original Redmine Issue: [19210](https://issues.qgis.org/issues/19210)
Affected QGIS version: 3.0.3
Redmine category:data_provider/postgis
---
When importing a layer that needed some schema tweaks (specifically to change the geom to a plain 'geometry' to avoid Polygon vs MultiPolygon errors... I wanted to turn off layer overwriting.
The ImportIntoPostGIS processing tool passes the overwrite option to
qgsvectorlayerexporter which passes it to
QgsPostgresProvider::createEmptyLayer
but as you can see here https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3878 and then https://github.com/qgis/QGIS/blob/0a325f1eda4054418e8d31bfc1d33d87cbdc2c0f/src/providers/postgres/qgspostgresprovider.cpp#L3893 the PostGIS driver only uses the overwrite flag to skip dropping the existing table (good), but not to avoid creating a new table (bad) - and this naturally fails.
This seems to be the case even very far back in history, so perhaps this should be considered a feature request rather than a bug?
|
process
|
allow importing data into existing tables with processing author name robert collins rbtcollins original redmine issue affected qgis version redmine category data provider postgis when importing a layer that needed some schema tweaks specifically to change the geom to a plain geometry to avoid polygon vs multipolygon errors i wanted to turn off layer overwriting the importintopostgis processing tool passes the overwrite option to qgsvectorlayerexporter which passes it to qgspostgresprovider createemptylayer but as you can see here and then the postgis driver only uses the overwrite flag to skip dropping the existing table good but not to avoid creating a new table bad and this naturally fails this seems to be the case even very far back in history so perhaps this should be considered a feature request rather than a bug
| 1
|
15,618
| 10,327,161,113
|
IssuesEvent
|
2019-09-02 05:53:40
|
vmware/singleton
|
https://api.github.com/repos/vmware/singleton
|
opened
|
[Singleton Service s3 build] cannot get the translation data from s3 manager, always return 520 error
|
area/service kind/bug priority/high
|
Singleton Service s3 build, all translation related APIs cannot return the correct translation data.
Product-based APIs
Component-based APIs
String-based APIs
The error message is:
```
{
"response": {
"code": 520,
"message": "1",
"serverTime": ""
},
"signature": "",
"data": ""
}
```
|
1.0
|
[Singleton Service s3 build] cannot get the translation data from s3 manager, always return 520 error - Singleton Service s3 build, all translation related APIs cannot return the correct translation data.
Product-based APIs
Component-based APIs
String-based APIs
The error message is:
```
{
"response": {
"code": 520,
"message": "1",
"serverTime": ""
},
"signature": "",
"data": ""
}
```
|
non_process
|
cannot get the translation data from manager always return error singleton service build all translation related apis cannot return the correct translation data product based apis component based apis string based apis the error message is response code message servertime signature data
| 0
|
108,674
| 13,645,951,382
|
IssuesEvent
|
2020-09-25 21:55:58
|
rubyforgood/casa
|
https://api.github.com/repos/rubyforgood/casa
|
closed
|
Add route to admins from sidebar menu
|
:paintbrush: Design :zap: Power :zap: Priority: High
|
Part of epic #822
As an admin,
I want to click on text "Admins" in the left sidebar,
and see all other admins appear in a contained wrapper,
so that I can take all actions associated with admins (create/view/edit) from the same place – without being directed to a different page.
**How does it work now?**
Admins can view a table of all other admins at the very bottom of their home dashboard. They display here by default, and don't go away.
**How should it work?**
Clicking on text "Admins" in left sidebar should prompt the admin view to appear in a contained wrapper on the same page.
**Acceptance criteria**
Text "Admins" appears in left sidebar
Clicking "Admins" opens the Admin view in a contained wrapper.
|
1.0
|
Add route to admins from sidebar menu - Part of epic #822
As an admin,
I want to click on text "Admins" in the left sidebar,
and see all other admins appear in a contained wrapper,
so that I can take all actions associated with admins (create/view/edit) from the same place – without being directed to a different page.
**How does it work now?**
Admins can view a table of all other admins at the very bottom of their home dashboard. They display here by default, and don't go away.
**How should it work?**
Clicking on text "Admins" in left sidebar should prompt the admin view to appear in a contained wrapper on the same page.
**Acceptance criteria**
Text "Admins" appears in left sidebar
Clicking "Admins" opens the Admin view in a contained wrapper.
|
non_process
|
add route to admins from sidebar menu part of epic as an admin i want to click on text admins in the left sidebar and see all other admins appear in a contained wrapper so that i can take all actions associated with admins create view edit from the same place – without being directed to a different page how does it work now admins can view a table of all other admins at the very bottom of their home dashboard they display here by default and don t go away how should it work clicking on text admins in left sidebar should prompt the admin view to appear in a contained wrapper on the same page acceptance criteria text admins appears in left sidebar clicking admins opens the admin view in a contained wrapper
| 0
|
127,341
| 5,028,782,872
|
IssuesEvent
|
2016-12-15 19:15:02
|
cyberpwnn/GlacialRealms
|
https://api.github.com/repos/cyberpwnn/GlacialRealms
|
closed
|
Titles Null Pointer Exception
|
20 Minute ETA bug high priority titles
|
```
[14:07:02 INFO]: cyberpwn issued server command: /title
[14:07:02 ERROR]: null
org.bukkit.command.CommandException: Unhandled exception executing command 'title' in plugin Titles v3.3
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:46) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:141) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at org.bukkit.craftbukkit.v1_8_R3.CraftServer.dispatchCommand(CraftServer.java:641) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PlayerConnection.handleCommand(PlayerConnection.java:1162) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PlayerConnection.a(PlayerConnection.java:997) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:45) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:1) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PlayerConnectionUtils$1.run(SourceFile:13) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) [?:1.8.0_101]
at java.util.concurrent.FutureTask.run(Unknown Source) [?:1.8.0_101]
at net.minecraft.server.v1_8_R3.SystemUtils.a(SourceFile:44) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:715) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:374) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:654) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:557) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at java.lang.Thread.run(Unknown Source) [?:1.8.0_101]
Caused by: java.lang.NullPointerException
at org.cyberpwn.titles.TitleController.hasTitles(TitleController.java:58) ~[?:?]
at org.cyberpwn.titles.CommandController.onCommand(CommandController.java:231) ~[?:?]
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:44) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
... 15 more
```
In the event where the following conditions are correct, npe's are thrown
* Player has NO data (missing keys)
* Phayer was online during a hotload of titles
* Player was already loaded in PDX
Solution: Create empty but existing keys before returning cluster
|
1.0
|
Titles Null Pointer Exception - ```
[14:07:02 INFO]: cyberpwn issued server command: /title
[14:07:02 ERROR]: null
org.bukkit.command.CommandException: Unhandled exception executing command 'title' in plugin Titles v3.3
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:46) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:141) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at org.bukkit.craftbukkit.v1_8_R3.CraftServer.dispatchCommand(CraftServer.java:641) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PlayerConnection.handleCommand(PlayerConnection.java:1162) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PlayerConnection.a(PlayerConnection.java:997) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:45) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PacketPlayInChat.a(PacketPlayInChat.java:1) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.PlayerConnectionUtils$1.run(SourceFile:13) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) [?:1.8.0_101]
at java.util.concurrent.FutureTask.run(Unknown Source) [?:1.8.0_101]
at net.minecraft.server.v1_8_R3.SystemUtils.a(SourceFile:44) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:715) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:374) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:654) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:557) [spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
at java.lang.Thread.run(Unknown Source) [?:1.8.0_101]
Caused by: java.lang.NullPointerException
at org.cyberpwn.titles.TitleController.hasTitles(TitleController.java:58) ~[?:?]
at org.cyberpwn.titles.CommandController.onCommand(CommandController.java:231) ~[?:?]
at org.bukkit.command.PluginCommand.execute(PluginCommand.java:44) ~[spigot-1.8.8.jar:git-Spigot-db6de12-18fbb24]
... 15 more
```
In the event where the following conditions are correct, npe's are thrown
* Player has NO data (missing keys)
* Phayer was online during a hotload of titles
* Player was already loaded in PDX
Solution: Create empty but existing keys before returning cluster
|
non_process
|
titles null pointer exception cyberpwn issued server command title null org bukkit command commandexception unhandled exception executing command title in plugin titles at org bukkit command plugincommand execute plugincommand java at org bukkit command simplecommandmap dispatch simplecommandmap java at org bukkit craftbukkit craftserver dispatchcommand craftserver java at net minecraft server playerconnection handlecommand playerconnection java at net minecraft server playerconnection a playerconnection java at net minecraft server packetplayinchat a packetplayinchat java at net minecraft server packetplayinchat a packetplayinchat java at net minecraft server playerconnectionutils run sourcefile at java util concurrent executors runnableadapter call unknown source at java util concurrent futuretask run unknown source at net minecraft server systemutils a sourcefile at net minecraft server minecraftserver b minecraftserver java at net minecraft server dedicatedserver b dedicatedserver java at net minecraft server minecraftserver a minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run unknown source caused by java lang nullpointerexception at org cyberpwn titles titlecontroller hastitles titlecontroller java at org cyberpwn titles commandcontroller oncommand commandcontroller java at org bukkit command plugincommand execute plugincommand java more in the event where the following conditions are correct npe s are thrown player has no data missing keys phayer was online during a hotload of titles player was already loaded in pdx solution create empty but existing keys before returning cluster
| 0
|
382,374
| 11,305,299,167
|
IssuesEvent
|
2020-01-18 04:11:47
|
siddhi-io/siddhi-io-http
|
https://api.github.com/repos/siddhi-io/siddhi-io-http
|
opened
|
Error while notifying listeners and request type is not a type of POST
|
priority/high type/bug
|
**Description:**
For the following Siddhi app deployment in K8s, I got the following error and the Siddhi runner did not come to the healthy state due to this error.
#### Siddhi App
```sql
@App:name("0.InputRoute-passthrough")
@source(type='http', receiver.url='http://0.0.0.0:9090/route', @map(type='json'))
define stream passthroughInputRouteStream(routeNo string);
@sink(type='nats',cluster.id='siddhi-stan',destination = '0.InputRoute_InputRouteStream', bootstrap.servers='nats://siddhi-nats:4222',@map(type='text'))
define stream InputRouteStream(routeNo string);
from passthroughInputRouteStream select * insert into InputRouteStream;
```
#### Error Log
```sh
[2020-01-18 02:50:47,196] ERROR {org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders} - Error while notifying listeners io.siddhi.extension.io.http.source.exception.HttpSourceAdaptorRuntimeException: Request type is not a type of POST
at io.siddhi.extension.io.http.source.HTTPConnectorListener.onMessage(HTTPConnectorListener.java:76)
at org.wso2.transport.http.netty.contractimpl.HttpWsServerConnectorFuture.notifyHttpListener(HttpWsServerConnectorFuture.java:72)
at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.notifyRequestListener(ReceivingHeaders.java:98)
at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.readInboundRequestHeaders(ReceivingHeaders.java:77)
at org.wso2.transport.http.netty.contractimpl.listener.states.ListenerReqRespStateManager.readInboundRequestHeaders(ListenerReqRespStateManager.java:38)
at org.wso2.transport.http.netty.contractimpl.listener.SourceHandler.channelRead(SourceHandler.java:130)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at org.wso2.transport.http.netty.contractimpl.listener.WebSocketServerHandshakeHandler.channelRead(WebSocketServerHandshakeHandler.java:135)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at org.wso2.transport.http.netty.contractimpl.listener.UriAndHeaderLengthValidator.channelRead(UriAndHeaderLengthValidator.java:64)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:310)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:284)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1359)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:935)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:141)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:645)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459)
at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:886)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Unknown Source)
```
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
Siddhi runner 5.1.2
|
1.0
|
Error while notifying listeners and request type is not a type of POST - **Description:**
For the following Siddhi app deployment in K8s, I got the following error and the Siddhi runner did not come to the healthy state due to this error.
#### Siddhi App
```sql
@App:name("0.InputRoute-passthrough")
@source(type='http', receiver.url='http://0.0.0.0:9090/route', @map(type='json'))
define stream passthroughInputRouteStream(routeNo string);
@sink(type='nats',cluster.id='siddhi-stan',destination = '0.InputRoute_InputRouteStream', bootstrap.servers='nats://siddhi-nats:4222',@map(type='text'))
define stream InputRouteStream(routeNo string);
from passthroughInputRouteStream select * insert into InputRouteStream;
```
#### Error Log
```sh
[2020-01-18 02:50:47,196] ERROR {org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders} - Error while notifying listeners io.siddhi.extension.io.http.source.exception.HttpSourceAdaptorRuntimeException: Request type is not a type of POST
at io.siddhi.extension.io.http.source.HTTPConnectorListener.onMessage(HTTPConnectorListener.java:76)
at org.wso2.transport.http.netty.contractimpl.HttpWsServerConnectorFuture.notifyHttpListener(HttpWsServerConnectorFuture.java:72)
at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.notifyRequestListener(ReceivingHeaders.java:98)
at org.wso2.transport.http.netty.contractimpl.listener.states.ReceivingHeaders.readInboundRequestHeaders(ReceivingHeaders.java:77)
at org.wso2.transport.http.netty.contractimpl.listener.states.ListenerReqRespStateManager.readInboundRequestHeaders(ListenerReqRespStateManager.java:38)
at org.wso2.transport.http.netty.contractimpl.listener.SourceHandler.channelRead(SourceHandler.java:130)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at org.wso2.transport.http.netty.contractimpl.listener.WebSocketServerHandshakeHandler.channelRead(WebSocketServerHandshakeHandler.java:135)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at org.wso2.transport.http.netty.contractimpl.listener.UriAndHeaderLengthValidator.channelRead(UriAndHeaderLengthValidator.java:64)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:310)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:284)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1359)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:935)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:141)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:645)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459)
at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:886)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Unknown Source)
```
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
Siddhi runner 5.1.2
|
non_process
|
error while notifying listeners and request type is not a type of post description for the following siddhi app deployment in i got the following error and the siddhi runner did not come to the healthy state due to this error siddhi app sql app name inputroute passthrough source type http receiver url map type json define stream passthroughinputroutestream routeno string sink type nats cluster id siddhi stan destination inputroute inputroutestream bootstrap servers nats siddhi nats map type text define stream inputroutestream routeno string from passthroughinputroutestream select insert into inputroutestream error log sh error org transport http netty contractimpl listener states receivingheaders error while notifying listeners io siddhi extension io http source exception httpsourceadaptorruntimeexception request type is not a type of post at io siddhi extension io http source httpconnectorlistener onmessage httpconnectorlistener java at org transport http netty contractimpl httpwsserverconnectorfuture notifyhttplistener httpwsserverconnectorfuture java at org transport http netty contractimpl listener states receivingheaders notifyrequestlistener receivingheaders java at org transport http netty contractimpl listener states receivingheaders readinboundrequestheaders receivingheaders java at org transport http netty contractimpl listener states listenerreqrespstatemanager readinboundrequestheaders listenerreqrespstatemanager java at org transport http netty contractimpl listener sourcehandler channelread sourcehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler timeout idlestatehandler channelread idlestatehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at org transport http netty contractimpl listener websocketserverhandshakehandler channelread websocketserverhandshakehandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at org transport http netty contractimpl listener uriandheaderlengthvalidator channelread uriandheaderlengthvalidator java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java lang thread run unknown source suggested labels suggested assignees affected product version siddhi runner
| 0
|
43,750
| 11,301,487,297
|
IssuesEvent
|
2020-01-17 15:41:32
|
AObuchow/lsp4xml-extensions-maven
|
https://api.github.com/repos/AObuchow/lsp4xml-extensions-maven
|
closed
|
Automated build + deploy
|
build
|
To improve feedback loop in m2e, it'd be great to have some automated builds available which would allow to easily include the latest build from master into m2e at build time (later, when we're ready, we'll probably stick to releases for more predictability/reproducibility but now, master is best).
To achieve so, best is probably to use GitHub Actions if possible.
|
1.0
|
Automated build + deploy - To improve feedback loop in m2e, it'd be great to have some automated builds available which would allow to easily include the latest build from master into m2e at build time (later, when we're ready, we'll probably stick to releases for more predictability/reproducibility but now, master is best).
To achieve so, best is probably to use GitHub Actions if possible.
|
non_process
|
automated build deploy to improve feedback loop in it d be great to have some automated builds available which would allow to easily include the latest build from master into at build time later when we re ready we ll probably stick to releases for more predictability reproducibility but now master is best to achieve so best is probably to use github actions if possible
| 0
|
15,236
| 19,141,678,157
|
IssuesEvent
|
2021-12-02 00:03:05
|
googleapis/repo-automation-bots
|
https://api.github.com/repos/googleapis/repo-automation-bots
|
closed
|
process(auto-approve): eventually we should phase out the json validation schema
|
type: process bot: auto approve
|
[This schema](https://github.com/googleapis/repo-automation-bots/blob/master/packages/auto-approve/src/valid-pr-schema.json) will soon become unwieldy with more PR use cases we acquire. We should start thinking about how to decide which PRs should be automerged in a smarter way.
|
1.0
|
process(auto-approve): eventually we should phase out the json validation schema - [This schema](https://github.com/googleapis/repo-automation-bots/blob/master/packages/auto-approve/src/valid-pr-schema.json) will soon become unwieldy with more PR use cases we acquire. We should start thinking about how to decide which PRs should be automerged in a smarter way.
|
process
|
process auto approve eventually we should phase out the json validation schema will soon become unwieldy with more pr use cases we acquire we should start thinking about how to decide which prs should be automerged in a smarter way
| 1
|
379,571
| 26,375,894,643
|
IssuesEvent
|
2023-01-12 02:23:37
|
cds-snc/resources-ressources
|
https://api.github.com/repos/cds-snc/resources-ressources
|
closed
|
Refactor HTML divider comments
|
Documentation Help wanted
|
It seems like Github does not like the following comment format
`<!-- [Text] ---------->`
Please reformat all HTML divider comments to look like the following:
`<!-- [Text] =========-->`
**Note:** This is relatively easy, and a good starter issue.
|
1.0
|
Refactor HTML divider comments - It seems like Github does not like the following comment format
`<!-- [Text] ---------->`
Please reformat all HTML divider comments to look like the following:
`<!-- [Text] =========-->`
**Note:** This is relatively easy, and a good starter issue.
|
non_process
|
refactor html divider comments it seems like github does not like the following comment format please reformat all html divider comments to look like the following note this is relatively easy and a good starter issue
| 0
|
16,628
| 21,701,781,117
|
IssuesEvent
|
2022-05-10 05:26:18
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Puppet Pals the Movie in 3-D
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Movie Puppet Pals the Movie in 3-D
Type (film/tv show): film
Film or show in which it appears: The Powerpuff Girls
Is the parent film/show streaming anywhere? HBO Max
About when in the parent film/show does it appear? yes
Actual footage of the film/show can be seen (yes/no)? yes
|
1.0
|
Add Puppet Pals the Movie in 3-D - Please add as much of the following info as you can:
Title: Movie Puppet Pals the Movie in 3-D
Type (film/tv show): film
Film or show in which it appears: The Powerpuff Girls
Is the parent film/show streaming anywhere? HBO Max
About when in the parent film/show does it appear? yes
Actual footage of the film/show can be seen (yes/no)? yes
|
process
|
add puppet pals the movie in d please add as much of the following info as you can title movie puppet pals the movie in d type film tv show film film or show in which it appears the powerpuff girls is the parent film show streaming anywhere hbo max about when in the parent film show does it appear yes actual footage of the film show can be seen yes no yes
| 1
|
8,795
| 11,908,210,241
|
IssuesEvent
|
2020-03-31 00:17:19
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Processing OGR based tools: encoding changed in outputs
|
Bug Feedback High Priority Processing Regression
|
Author Name: **Bernd Vogelgesang** (Bernd Vogelgesang)
Original Redmine Issue: [21949](https://issues.qgis.org/issues/21949)
Affected QGIS version: 3.6.2
Redmine category:processing/ogr
---
When trying to buffer a point dataset with the GDAL buffer vector function, the attributes get obviously encoded to sth else than the original UTF-8.
All special characters become a ? on black rhombic background. When manually setting the resulting layer to windows1252, the special characters are visible again.
This seems to apply to other GDAL/OGR functions besides buffer as well.
Besides that I would like to second Tobias's feature request #29097 to change the output format from ESRI shapefile to GPKG, cause shape format (dbase) might cripple any GPKG-based workflow. And as GPKG is UTF-8 by default, the encoding issues would/could/should be gone as well, I hope.
QGIS 3.6.2 on Linux Mint
|
1.0
|
Processing OGR based tools: encoding changed in outputs - Author Name: **Bernd Vogelgesang** (Bernd Vogelgesang)
Original Redmine Issue: [21949](https://issues.qgis.org/issues/21949)
Affected QGIS version: 3.6.2
Redmine category:processing/ogr
---
When trying to buffer a point dataset with the GDAL buffer vector function, the attributes get obviously encoded to sth else than the original UTF-8.
All special characters become a ? on black rhombic background. When manually setting the resulting layer to windows1252, the special characters are visible again.
This seems to apply to other GDAL/OGR functions besides buffer as well.
Besides that I would like to second Tobias's feature request #29097 to change the output format from ESRI shapefile to GPKG, cause shape format (dbase) might cripple any GPKG-based workflow. And as GPKG is UTF-8 by default, the encoding issues would/could/should be gone as well, I hope.
QGIS 3.6.2 on Linux Mint
|
process
|
processing ogr based tools encoding changed in outputs author name bernd vogelgesang bernd vogelgesang original redmine issue affected qgis version redmine category processing ogr when trying to buffer a point dataset with the gdal buffer vector function the attributes get obviously encoded to sth else than the original utf all special characters become a on black rhombic background when manually setting the resulting layer to the special characters are visible again this seems to apply to other gdal ogr functions besides buffer as well besides that i would like to second tobias s feature request to change the output format from esri shapefile to gpkg cause shape format dbase might cripple any gpkg based workflow and as gpkg is utf by default the encoding issues would could should be gone as well i hope qgis on linux mint
| 1
|
304,282
| 26,265,440,050
|
IssuesEvent
|
2023-01-06 12:05:02
|
NationalSecurityAgency/skills-service
|
https://api.github.com/repos/NationalSecurityAgency/skills-service
|
closed
|
Utilize local storage for an extra level of resiliency to preserve skill/badge/project/subject attributes (ex. description)
|
enhancement test
|
Add an extra level of protection when users spend time to create a **new** skill/badge/project/subject attributes followed by the failure to save those changes (due to the networking hiccup, infrastructure instability or a SkillTree bug). Prevent entered information from being lost in the event of unexpected failures.
- Store attributes in local storage as they are entered
- clear local storage when information is saved or modal is closed by the user
- (optional): when closing a modal warn users if the values changed since the modal was opened
- keep the information in local storage unless modal is closed by the user which addresses application errors, networking hiccups, accidental refresh, computer restarts, etc...
- alert the user if the information is restored
|
1.0
|
Utilize local storage for an extra level of resiliency to preserve skill/badge/project/subject attributes (ex. description) - Add an extra level of protection when users spend time to create a **new** skill/badge/project/subject attributes followed by the failure to save those changes (due to the networking hiccup, infrastructure instability or a SkillTree bug). Prevent entered information from being lost in the event of unexpected failures.
- Store attributes in local storage as they are entered
- clear local storage when information is saved or modal is closed by the user
- (optional): when closing a modal warn users if the values changed since the modal was opened
- keep the information in local storage unless modal is closed by the user which addresses application errors, networking hiccups, accidental refresh, computer restarts, etc...
- alert the user if the information is restored
|
non_process
|
utilize local storage for an extra level of resiliency to preserve skill badge project subject attributes ex description add an extra level of protection when users spend time to create a new skill badge project subject attributes followed by the failure to save those changes due to the networking hiccup infrastructure instability or a skilltree bug prevent entered information from being lost in the event of unexpected failures store attributes in local storage as they are entered clear local storage when information is saved or modal is closed by the user optional when closing a modal warn users if the values changed since the modal was opened keep the information in local storage unless modal is closed by the user which addresses application errors networking hiccups accidental refresh computer restarts etc alert the user if the information is restored
| 0
|
22,908
| 3,727,389,446
|
IssuesEvent
|
2016-03-06 08:05:05
|
godfather1103/mentohust
|
https://api.github.com/repos/godfather1103/mentohust
|
closed
|
HUST昨天今天一直说版本过低,小女子求救!
|
auto-migrated Priority-Medium Type-Defect
|
```
:: 欢迎使用MentoHUST!
** 本机MAC: e8-03-9a-39-37-0b
** 本机IP: 172.20.86.43
** 子网掩码: 255.255.252.0
** 默认网关; 172.20.84.1
** 首选DNS: 202.116.64.107
>> 寻找服务器...
** 认证MAC: 00-1a-a9-17-ff-ff
>> 发送用户名...
>> 发送密码...
>> 认证失败!
$$ 系统提示: 锐捷客户端版本过低!A?;牖?u?哠?糛嘾
>> 等待服务器请求或15秒后重启认证。
** 网关MAC: 00-d0-f8-d5-a8-52
** 认证MAC: 00-1a-a9-17-ff-ff
>> 发送用户名...
>> 寻找服务器...
>> 发送密码...
>> 发送用户名...
>> 发送密码...
>> 认证失败!
$$ 系统提示: 锐捷客户端版本过低!A兖K?漥1?n?糛Om
>> 等待服务器请求或15秒后重启认证。
:: 认证已断开!
```
Original issue reported on code.google.com by `newh...@gmail.com` on 15 Jun 2013 at 10:18
|
1.0
|
HUST昨天今天一直说版本过低,小女子求救! - ```
:: 欢迎使用MentoHUST!
** 本机MAC: e8-03-9a-39-37-0b
** 本机IP: 172.20.86.43
** 子网掩码: 255.255.252.0
** 默认网关; 172.20.84.1
** 首选DNS: 202.116.64.107
>> 寻找服务器...
** 认证MAC: 00-1a-a9-17-ff-ff
>> 发送用户名...
>> 发送密码...
>> 认证失败!
$$ 系统提示: 锐捷客户端版本过低!A?;牖?u?哠?糛嘾
>> 等待服务器请求或15秒后重启认证。
** 网关MAC: 00-d0-f8-d5-a8-52
** 认证MAC: 00-1a-a9-17-ff-ff
>> 发送用户名...
>> 寻找服务器...
>> 发送密码...
>> 发送用户名...
>> 发送密码...
>> 认证失败!
$$ 系统提示: 锐捷客户端版本过低!A兖K?漥1?n?糛Om
>> 等待服务器请求或15秒后重启认证。
:: 认证已断开!
```
Original issue reported on code.google.com by `newh...@gmail.com` on 15 Jun 2013 at 10:18
|
non_process
|
hust昨天今天一直说版本过低,小女子求救! 欢迎使用mentohust 本机mac 本机ip 子网掩码 默认网关 首选dns 寻找服务器 认证mac ff ff 发送用户名 发送密码 认证失败 系统提示 锐捷客户端版本过低 a 牖 u 哠 糛嘾 。 网关mac 认证mac ff ff 发送用户名 寻找服务器 发送密码 发送用户名 发送密码 认证失败 系统提示 锐捷客户端版本过低 a兖k n 糛om 。 认证已断开 original issue reported on code google com by newh gmail com on jun at
| 0
|
9,087
| 12,154,523,967
|
IssuesEvent
|
2020-04-25 08:45:21
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
opened
|
`ess-set-working-directory` should not change `default-directory`
|
process
|
I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R.
It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the file I'm switching from.
Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
|
1.0
|
`ess-set-working-directory` should not change `default-directory` - I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R.
It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the file I'm switching from.
Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
|
process
|
ess set working directory should not change default directory i ve been confused for years with the behaviour of emacs regarding the active directories of the files i m editing i just realised it s ess set working directory that changes it to the new working directory of r it seems that this behaviour is intended but i m not sure why changing the repl directory should affect the file i m switching from should we change this to preserve the active directory the active directory influences ui commands like m x find file i suspect i m not the only one who is confused by this behaviour
| 1
|
7,735
| 10,854,882,948
|
IssuesEvent
|
2019-11-13 17:13:25
|
codeuniversity/smag-mvp
|
https://api.github.com/repos/codeuniversity/smag-mvp
|
opened
|
Get user data from graphdb using face recognition result
|
Backend Image Processing
|
After uploading and analysing the face picture, the data of the user has to be retrieved from graphdb using the encoding.
|
1.0
|
Get user data from graphdb using face recognition result - After uploading and analysing the face picture, the data of the user has to be retrieved from graphdb using the encoding.
|
process
|
get user data from graphdb using face recognition result after uploading and analysing the face picture the data of the user has to be retrieved from graphdb using the encoding
| 1
|
16,155
| 20,515,161,753
|
IssuesEvent
|
2022-03-01 10:57:37
|
decidim/decidim
|
https://api.github.com/repos/decidim/decidim
|
closed
|
Show statistics and show metrics are always enabled when creating a new process (ignoring admin selection)
|
type: bug module: participatory processes
|
**Describe the bug**
When I create a new process, "show statistics" and "show metrics" are always enabled regardless I have selected them or not.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Processes and Create a New Process
2. Keep unselected "show statistics", "show metrics" or both
3. Create the new process
4. Review process info > "show statistics" and "show metrics" are both selected
**Expected behavior**
If "show statistics" or "show metrics" are not selected, keep them unselected when creating the process.
**Screenshots**
https://user-images.githubusercontent.com/72607737/155357502-9c97f784-814d-446e-a730-3e2d9893d391.mov
**Extra data (please complete the following information):**
- Device: Desktop
- Device OS: macOS Catalina
- Browser: Chrome & Firefox
- Decidim Version: 0.25.2
- Decidim installation: staging.platoniq.net
|
1.0
|
Show statistics and show metrics are always enabled when creating a new process (ignoring admin selection) - **Describe the bug**
When I create a new process, "show statistics" and "show metrics" are always enabled regardless I have selected them or not.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to Processes and Create a New Process
2. Keep unselected "show statistics", "show metrics" or both
3. Create the new process
4. Review process info > "show statistics" and "show metrics" are both selected
**Expected behavior**
If "show statistics" or "show metrics" are not selected, keep them unselected when creating the process.
**Screenshots**
https://user-images.githubusercontent.com/72607737/155357502-9c97f784-814d-446e-a730-3e2d9893d391.mov
**Extra data (please complete the following information):**
- Device: Desktop
- Device OS: macOS Catalina
- Browser: Chrome & Firefox
- Decidim Version: 0.25.2
- Decidim installation: staging.platoniq.net
|
process
|
show statistics and show metrics are always enabled when creating a new process ignoring admin selection describe the bug when i create a new process show statistics and show metrics are always enabled regardless i have selected them or not to reproduce steps to reproduce the behavior go to processes and create a new process keep unselected show statistics show metrics or both create the new process review process info show statistics and show metrics are both selected expected behavior if show statistics or show metrics are not selected keep them unselected when creating the process screenshots extra data please complete the following information device desktop device os macos catalina browser chrome firefox decidim version decidim installation staging platoniq net
| 1
|
11,095
| 13,937,850,379
|
IssuesEvent
|
2020-10-22 14:36:15
|
prisma/prisma-engines
|
https://api.github.com/repos/prisma/prisma-engines
|
opened
|
Introduce support for scalar type `BigInt`
|
engines/data model parser engines/introspection engine engines/migration engine engines/query engine process/candidate team/engines
|
We decided to introduce the scalar type `BigInt`.
So that means that the following schema is not valid anymore:
```
model Blog {
id Int @id
bigInt Int @mydb.BigInt
}
```
Instead users will have to write it like this:
```
model Blog {
id Int @id
bigInt BigInt
}
```
|
1.0
|
Introduce support for scalar type `BigInt` - We decided to introduce the scalar type `BigInt`.
So that means that the following schema is not valid anymore:
```
model Blog {
id Int @id
bigInt Int @mydb.BigInt
}
```
Instead users will have to write it like this:
```
model Blog {
id Int @id
bigInt BigInt
}
```
|
process
|
introduce support for scalar type bigint we decided to introduce the scalar type bigint so that means that the following schema is not valid anymore model blog id int id bigint int mydb bigint instead users will have to write it like this model blog id int id bigint bigint
| 1
|
2,216
| 11,592,816,185
|
IssuesEvent
|
2020-02-24 12:20:08
|
big-neon/bn-web
|
https://api.github.com/repos/big-neon/bn-web
|
opened
|
Automation: Big Neon: Test: Allow Past Events to Remain On Site: Search Using URL
|
Automation
|
**Pre-conditions:**
1. User should have admin access to Big Neon
2. User should be logged into Big Neon
3. User should have an event that has past
**Steps:**
1. Add the URL to view the event that has past
2. View event page loads successfully
3. Verify the button "Purchase Tickets" is now displayed as "This Event Is Now Over" with a tear emoji
4. Try to select the above button
5. User should be unable to select the button
test pad link: https://big-neon.ontestpad.com/script/194#11//
|
1.0
|
Automation: Big Neon: Test: Allow Past Events to Remain On Site: Search Using URL - **Pre-conditions:**
1. User should have admin access to Big Neon
2. User should be logged into Big Neon
3. User should have an event that has past
**Steps:**
1. Add the URL to view the event that has past
2. View event page loads successfully
3. Verify the button "Purchase Tickets" is now displayed as "This Event Is Now Over" with a tear emoji
4. Try to select the above button
5. User should be unable to select the button
test pad link: https://big-neon.ontestpad.com/script/194#11//
|
non_process
|
automation big neon test allow past events to remain on site search using url pre conditions user should have admin access to big neon user should be logged into big neon user should have an event that has past steps add the url to view the event that has past view event page loads successfully verify the button purchase tickets is now displayed as this event is now over with a tear emoji try to select the above button user should be unable to select the button test pad link
| 0
|
303,863
| 9,311,343,138
|
IssuesEvent
|
2019-03-25 21:07:08
|
ngageoint/hootenanny
|
https://api.github.com/repos/ngageoint/hootenanny
|
closed
|
Not able to find merger for match combination when conflating Bandug data
|
Category: Algorithms Priority: Medium Status: Defined Type: Bug
|
Must run with network conflation for this (see command below). This is the follow up task to #2059 which put a bandaid on the problem. It is possible that the fixes for #2059 are valid by themselves, but that needs to be proven as part of this task.
Input data:
ref: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/MapEditBandug.osm
sec: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/OSMmap.osm
If you re-enable the exception that used to be thrown in MergerFactory, you will see that a NetworkMatch and a ScriptMatch are grouped together to be merged, which I don't think makes sense. Since no merger can merge those two, the exception ends up being thrown. If this is proven to be a bug and gets fixed, you can also remove the extra error checking that was added to ScriptMerger::_applyMergePair as part of #2059.
`hoot crop-map MapEditBandug.osm MapEditBandug-cropped.osm "107.604,-6.922,107.611,-6.915"`
`hoot crop-map OSMmap.osm OSMmap-cropped.osm "107.604,-6.922,107.611,-6.915"`
`hoot conflate -D uuid.helper.repeatable=true -D writer.include.debug.tags=true -D match.creators="hoot::BuildingMatchCreator;hoot::ScriptMatchCreator,PoiGeneric.js;hoot::NetworkMatchCreator;hoot::ScriptMatchCreator,LinearWaterway.js;hoot::PoiPolygonMatchCreator;hoot::ScriptMatchCreator,Area.js" -D merger.creators="hoot::BuildingMergerCreator;hoot::ScriptMergerCreator;hoot::NetworkMergerCreator;hoot::PoiPolygonMergerCreator;hoot::ScriptMergerCreator" -D conflate.match.highway.classifier="hoot::HighwayExpertClassifier" -D way.subline.matcher="hoot::MaximalSublineMatcher" -D rubber.sheet.minimum.ties=4 -D rubber.sheet.ref=true MapEditBandug-cropped.osm OSMmap-cropped.osm out.osm`
|
1.0
|
Not able to find merger for match combination when conflating Bandug data - Must run with network conflation for this (see command below). This is the follow up task to #2059 which put a bandaid on the problem. It is possible that the fixes for #2059 are valid by themselves, but that needs to be proven as part of this task.
Input data:
ref: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/MapEditBandug.osm
sec: https://github.com/DigitalGlobe/VGI-team-repo/blob/HootenannyConflation/OSMmap.osm
If you re-enable the exception that used to be thrown in MergerFactory, you will see that a NetworkMatch and a ScriptMatch are grouped together to be merged, which I don't think makes sense. Since no merger can merge those two, the exception ends up being thrown. If this is proven to be a bug and gets fixed, you can also remove the extra error checking that was added to ScriptMerger::_applyMergePair as part of #2059.
`hoot crop-map MapEditBandug.osm MapEditBandug-cropped.osm "107.604,-6.922,107.611,-6.915"`
`hoot crop-map OSMmap.osm OSMmap-cropped.osm "107.604,-6.922,107.611,-6.915"`
`hoot conflate -D uuid.helper.repeatable=true -D writer.include.debug.tags=true -D match.creators="hoot::BuildingMatchCreator;hoot::ScriptMatchCreator,PoiGeneric.js;hoot::NetworkMatchCreator;hoot::ScriptMatchCreator,LinearWaterway.js;hoot::PoiPolygonMatchCreator;hoot::ScriptMatchCreator,Area.js" -D merger.creators="hoot::BuildingMergerCreator;hoot::ScriptMergerCreator;hoot::NetworkMergerCreator;hoot::PoiPolygonMergerCreator;hoot::ScriptMergerCreator" -D conflate.match.highway.classifier="hoot::HighwayExpertClassifier" -D way.subline.matcher="hoot::MaximalSublineMatcher" -D rubber.sheet.minimum.ties=4 -D rubber.sheet.ref=true MapEditBandug-cropped.osm OSMmap-cropped.osm out.osm`
|
non_process
|
not able to find merger for match combination when conflating bandug data must run with network conflation for this see command below this is the follow up task to which put a bandaid on the problem it is possible that the fixes for are valid by themselves but that needs to be proven as part of this task input data ref sec if you re enable the exception that used to be thrown in mergerfactory you will see that a networkmatch and a scriptmatch are grouped together to be merged which i don t think makes sense since no merger can merge those two the exception ends up being thrown if this is proven to be a bug and gets fixed you can also remove the extra error checking that was added to scriptmerger applymergepair as part of hoot crop map mapeditbandug osm mapeditbandug cropped osm hoot crop map osmmap osm osmmap cropped osm hoot conflate d uuid helper repeatable true d writer include debug tags true d match creators hoot buildingmatchcreator hoot scriptmatchcreator poigeneric js hoot networkmatchcreator hoot scriptmatchcreator linearwaterway js hoot poipolygonmatchcreator hoot scriptmatchcreator area js d merger creators hoot buildingmergercreator hoot scriptmergercreator hoot networkmergercreator hoot poipolygonmergercreator hoot scriptmergercreator d conflate match highway classifier hoot highwayexpertclassifier d way subline matcher hoot maximalsublinematcher d rubber sheet minimum ties d rubber sheet ref true mapeditbandug cropped osm osmmap cropped osm out osm
| 0
|
5,810
| 8,644,718,444
|
IssuesEvent
|
2018-11-26 04:38:01
|
gfrebello/qs-trip-planning-procedure
|
https://api.github.com/repos/gfrebello/qs-trip-planning-procedure
|
closed
|
Realize Tests and Update Test Document
|
Priority:High Process:Run Test Case
|
Tests need to be realized to make sure the requirements are being met, and if not, either the project needs to be corrected, or the requirements need to be changed.
|
1.0
|
Realize Tests and Update Test Document - Tests need to be realized to make sure the requirements are being met, and if not, either the project needs to be corrected, or the requirements need to be changed.
|
process
|
realize tests and update test document tests need to be realized to make sure the requirements are being met and if not either the project needs to be corrected or the requirements need to be changed
| 1
|
234,161
| 19,098,621,353
|
IssuesEvent
|
2021-11-29 19:34:20
|
microsoft/vscode-python
|
https://api.github.com/repos/microsoft/vscode-python
|
opened
|
Test Pylance with a virtual workspace
|
testplan-item
|
Refs: #17253
- [ ] anyOS
- [ ] anyOS
- [ ] anyOS
Complexity: 2
Authors: @karrtikr
---
### Requirements
1. Install python extension from here: https://pvsc.blob.core.windows.net/extension-builds/ms-python-insiders.vsix
### Verification
1. Open a GitHub repository containing Python code, and make sure it is trusted
2. Make sure that you have the `python.languageServer` setting set to `Default` or `Pylance`
3. Activate the Python extension
4. Verify that the `Starting Pylance language server.` message appears in the Python output channel
5. Make sure you get hover info, same-file completion and completion for built-ins like `print`
|
1.0
|
Test Pylance with a virtual workspace - Refs: #17253
- [ ] anyOS
- [ ] anyOS
- [ ] anyOS
Complexity: 2
Authors: @karrtikr
---
### Requirements
1. Install python extension from here: https://pvsc.blob.core.windows.net/extension-builds/ms-python-insiders.vsix
### Verification
1. Open a GitHub repository containing Python code, and make sure it is trusted
2. Make sure that you have the `python.languageServer` setting set to `Default` or `Pylance`
3. Activate the Python extension
4. Verify that the `Starting Pylance language server.` message appears in the Python output channel
5. Make sure you get hover info, same-file completion and completion for built-ins like `print`
|
non_process
|
test pylance with a virtual workspace refs anyos anyos anyos complexity authors karrtikr requirements install python extension from here verification open a github repository containing python code and make sure it is trusted make sure that you have the python languageserver setting set to default or pylance activate the python extension verify that the starting pylance language server message appears in the python output channel make sure you get hover info same file completion and completion for built ins like print
| 0
|
268,479
| 23,373,633,030
|
IssuesEvent
|
2022-08-10 22:51:36
|
kubernetes/test-infra
|
https://api.github.com/repos/kubernetes/test-infra
|
closed
|
Job names have wrong suffix in new branch
|
kind/bug sig/testing sig/release
|
**What happened**:
When generating the new 1.25 branch jobs, the names of the jobs were wrongly generated, causing a whole bunch of presubmits to fail when verifying the config files.
The job names had the release version appended (`-1.25`)
**What you expected to happen**:
Job names should not have had the suffixes added
**How to reproduce it (as minimally and precisely as possible)**:
Ran `make -C releng prepare-release-branch`
**Please provide links to example occurrences, if any**:
https://github.com/kubernetes/test-infra/pull/27096/commits/8058610ea3dd366a2153f5d8a8bf0adfc81d39d9
**Anything else we need to know?**:
|
1.0
|
Job names have wrong suffix in new branch -
**What happened**:
When generating the new 1.25 branch jobs, the names of the jobs were wrongly generated, causing a whole bunch of presubmits to fail when verifying the config files.
The job names had the release version appended (`-1.25`)
**What you expected to happen**:
Job names should not have had the suffixes added
**How to reproduce it (as minimally and precisely as possible)**:
Ran `make -C releng prepare-release-branch`
**Please provide links to example occurrences, if any**:
https://github.com/kubernetes/test-infra/pull/27096/commits/8058610ea3dd366a2153f5d8a8bf0adfc81d39d9
**Anything else we need to know?**:
|
non_process
|
job names have wrong suffix in new branch what happened when generating the new branch jobs the names of the jobs were wrongly generated causing a whole bunch of presubmits to fail when verifying the config files the job names had the release version appended what you expected to happen job names should not have had the suffixes added how to reproduce it as minimally and precisely as possible ran make c releng prepare release branch please provide links to example occurrences if any anything else we need to know
| 0
|
16,279
| 20,884,554,008
|
IssuesEvent
|
2022-03-23 02:34:50
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add All Over The Girl
|
suggested title in process
|
Please add as much of the following info as you can:
Title: All Over The Girl
Type (film/tv show): Film
Film or show in which it appears: Marriage Story
Is the parent film/show streaming anywhere? [Netflix](https://www.netflix.com/title/80223779)
About when in the parent film/show does it appear? 2m50s
Actual footage of the film/show can be seen (yes/no)? Yes
|
1.0
|
Add All Over The Girl - Please add as much of the following info as you can:
Title: All Over The Girl
Type (film/tv show): Film
Film or show in which it appears: Marriage Story
Is the parent film/show streaming anywhere? [Netflix](https://www.netflix.com/title/80223779)
About when in the parent film/show does it appear? 2m50s
Actual footage of the film/show can be seen (yes/no)? Yes
|
process
|
add all over the girl please add as much of the following info as you can title all over the girl type film tv show film film or show in which it appears marriage story is the parent film show streaming anywhere about when in the parent film show does it appear actual footage of the film show can be seen yes no yes
| 1
|
661,559
| 22,060,677,611
|
IssuesEvent
|
2022-05-30 17:21:25
|
SimpleGeek/squid-config-ui
|
https://api.github.com/repos/SimpleGeek/squid-config-ui
|
opened
|
Bypass menu not correctly displaying time after key sequence
|
bug Priority: Medium
|
To reproduce:
Enter a valid time, then hit enter again. You should get the error alert that the time was invalid, and then the bypass list should say there are no active bypasses, despite a bypass being active for the time assigned in the first step.
|
1.0
|
Bypass menu not correctly displaying time after key sequence - To reproduce:
Enter a valid time, then hit enter again. You should get the error alert that the time was invalid, and then the bypass list should say there are no active bypasses, despite a bypass being active for the time assigned in the first step.
|
non_process
|
bypass menu not correctly displaying time after key sequence to reproduce enter a valid time then hit enter again you should get the error alert that the time was invalid and then the bypass list should say there are no active bypasses despite a bypass being active for the time assigned in the first step
| 0
|
19,794
| 26,178,260,419
|
IssuesEvent
|
2023-01-02 12:31:29
|
mdsreq-fga-unb/2022.2-Receitalista
|
https://api.github.com/repos/mdsreq-fga-unb/2022.2-Receitalista
|
closed
|
Colocar as atividades de cada evento Scrum
|
processo visao
|
No tópico 3.2 do documento da Visão, deve-se separar os eventos do Scrum para colocar as atividades de cada evento.
**Como está:**
|Atividade|Método|Ferramenta|Entrega|
|:-----:|:------:|:------:|:------:|
|Planejamento do Sprint (sprint planning)|Reunião|Discord|Backlog do Sprint|
|Revisão do Sprint (sprint review)|Reunião|Discord|Nova release|
|Reunião diária (daily)|Conversa entre os integrantes da equipe|Telegram|Feedback do Sprint|
|Retrospectiva do Sprint (sprint retrospecive)|Reunião|Discord|Feedback do Sprint|
**Como deve ser:**
### 3.n - Planejamento do Sprint (sprint planning)
|Atividade|Método|Ferramenta|Entrega|
|:-----:|:------:|:------:|:------:|
|Atividade 1|etc|etc|etc|
|Atividade 2|etc|etc|etc|
### 3.n - Revisão do Sprint (sprint review)
|Atividade|Método|Ferramenta|Entrega|
|:-----:|:------:|:------:|:------:|
|Atividade 1|etc|etc|etc|
|Atividade 2|etc|etc|etc|
|
1.0
|
Colocar as atividades de cada evento Scrum - No tópico 3.2 do documento da Visão, deve-se separar os eventos do Scrum para colocar as atividades de cada evento.
**Como está:**
|Atividade|Método|Ferramenta|Entrega|
|:-----:|:------:|:------:|:------:|
|Planejamento do Sprint (sprint planning)|Reunião|Discord|Backlog do Sprint|
|Revisão do Sprint (sprint review)|Reunião|Discord|Nova release|
|Reunião diária (daily)|Conversa entre os integrantes da equipe|Telegram|Feedback do Sprint|
|Retrospectiva do Sprint (sprint retrospecive)|Reunião|Discord|Feedback do Sprint|
**Como deve ser:**
### 3.n - Planejamento do Sprint (sprint planning)
|Atividade|Método|Ferramenta|Entrega|
|:-----:|:------:|:------:|:------:|
|Atividade 1|etc|etc|etc|
|Atividade 2|etc|etc|etc|
### 3.n - Revisão do Sprint (sprint review)
|Atividade|Método|Ferramenta|Entrega|
|:-----:|:------:|:------:|:------:|
|Atividade 1|etc|etc|etc|
|Atividade 2|etc|etc|etc|
|
process
|
colocar as atividades de cada evento scrum no tópico do documento da visão deve se separar os eventos do scrum para colocar as atividades de cada evento como está atividade método ferramenta entrega planejamento do sprint sprint planning reunião discord backlog do sprint revisão do sprint sprint review reunião discord nova release reunião diária daily conversa entre os integrantes da equipe telegram feedback do sprint retrospectiva do sprint sprint retrospecive reunião discord feedback do sprint como deve ser n planejamento do sprint sprint planning atividade método ferramenta entrega atividade etc etc etc atividade etc etc etc n revisão do sprint sprint review atividade método ferramenta entrega atividade etc etc etc atividade etc etc etc
| 1
|
1,112
| 9,488,823,580
|
IssuesEvent
|
2019-04-22 20:38:21
|
mozilla-mobile/reference-browser
|
https://api.github.com/repos/mozilla-mobile/reference-browser
|
closed
|
Dep-sign debug builds on-push on the master branch
|
🤖 automation
|
We will have performance tests running on `reference-browser` in the future. One of the requirements is to have builds signed with the dep key.
This ticket requires that, on every commit to the master branch, all the debug variants should be dep-signed.
|
1.0
|
Dep-sign debug builds on-push on the master branch - We will have performance tests running on `reference-browser` in the future. One of the requirements is to have builds signed with the dep key.
This ticket requires that, on every commit to the master branch, all the debug variants should be dep-signed.
|
non_process
|
dep sign debug builds on push on the master branch we will have performance tests running on reference browser in the future one of the requirements is to have builds signed with the dep key this ticket requires that on every commit to the master branch all the debug variants should be dep signed
| 0
|
3,774
| 4,043,821,506
|
IssuesEvent
|
2016-05-21 00:13:57
|
google/cadvisor
|
https://api.github.com/repos/google/cadvisor
|
closed
|
[scalability] failed to collect filesystem stats - failed to exec du
|
area/performance
|
Full error:
```
E0520 20:38:01.903710 16716 fsHandler.go:106] failed to collect filesystem stats - failed to exec du - fork/exec /usr/bin/nice: cannot allocate memory
```
To reproduce:
```
$ sudo ./cadvisor --logtostderr &
$ for i in `seq 1 500`; do docker run -d kubernetes/pause; done
```
I suggest we limit the number of parallel execs to a reasonable number.
|
True
|
[scalability] failed to collect filesystem stats - failed to exec du - Full error:
```
E0520 20:38:01.903710 16716 fsHandler.go:106] failed to collect filesystem stats - failed to exec du - fork/exec /usr/bin/nice: cannot allocate memory
```
To reproduce:
```
$ sudo ./cadvisor --logtostderr &
$ for i in `seq 1 500`; do docker run -d kubernetes/pause; done
```
I suggest we limit the number of parallel execs to a reasonable number.
|
non_process
|
failed to collect filesystem stats failed to exec du full error fshandler go failed to collect filesystem stats failed to exec du fork exec usr bin nice cannot allocate memory to reproduce sudo cadvisor logtostderr for i in seq do docker run d kubernetes pause done i suggest we limit the number of parallel execs to a reasonable number
| 0
|
10,306
| 13,155,329,080
|
IssuesEvent
|
2020-08-10 08:41:19
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
[Bug report] style scoped page选择器能否过滤
|
processing
|
**问题描述**
使用autoScopeRules时如果页面文件中包含页面,支付宝下页面中含有page选择器会也会加上scopedId,导致page选择器无效,scoped这里可以针对page选择器进行过滤吗?
**环境信息描述**
Mac
mpx版本
"@mpxjs/core": "^2.5.33"
"@mpxjs/webpack-plugin": "^2.5.33"
支付宝IDE 1.13.4
**最简复现demo**
```javascript
// mpx.webpack.conf.js
{
autoScopeRules: {
include: [resolve('src')],
exclude: [
resolve('src/app.mpx'),
resolve('src/assets'),
resolve('src/components/@vant'),
],
},
}
```
```css
// src/pages/xxx.mpx
page {
background-color: #f8f8f8
}
// 编译后
page.m9ed7a486 {
background-color: #f8f8f8;
}
```
|
1.0
|
[Bug report] style scoped page选择器能否过滤 - **问题描述**
使用autoScopeRules时如果页面文件中包含页面,支付宝下页面中含有page选择器会也会加上scopedId,导致page选择器无效,scoped这里可以针对page选择器进行过滤吗?
**环境信息描述**
Mac
mpx版本
"@mpxjs/core": "^2.5.33"
"@mpxjs/webpack-plugin": "^2.5.33"
支付宝IDE 1.13.4
**最简复现demo**
```javascript
// mpx.webpack.conf.js
{
autoScopeRules: {
include: [resolve('src')],
exclude: [
resolve('src/app.mpx'),
resolve('src/assets'),
resolve('src/components/@vant'),
],
},
}
```
```css
// src/pages/xxx.mpx
page {
background-color: #f8f8f8
}
// 编译后
page.m9ed7a486 {
background-color: #f8f8f8;
}
```
|
process
|
style scoped page选择器能否过滤 问题描述 使用autoscoperules时如果页面文件中包含页面,支付宝下页面中含有page选择器会也会加上scopedid,导致page选择器无效,scoped这里可以针对page选择器进行过滤吗? 环境信息描述 mac mpx版本 mpxjs core mpxjs webpack plugin 支付宝ide 最简复现demo javascript mpx webpack conf js autoscoperules include exclude resolve src app mpx resolve src assets resolve src components vant css src pages xxx mpx page background color 编译后 page background color
| 1
|
14,376
| 17,398,554,278
|
IssuesEvent
|
2021-08-02 16:17:15
|
pystatgen/sgkit
|
https://api.github.com/repos/pystatgen/sgkit
|
closed
|
Release 0.3.0
|
process + tools
|
This is a coordinating issue for the release.
Issues for this release are listed here: https://github.com/pystatgen/sgkit/milestone/1
The artifacts for the previous release (#495) were built manually. For this release I would like to build and deploy artifacts using GH actions, see #606 and #607.
|
1.0
|
Release 0.3.0 - This is a coordinating issue for the release.
Issues for this release are listed here: https://github.com/pystatgen/sgkit/milestone/1
The artifacts for the previous release (#495) were built manually. For this release I would like to build and deploy artifacts using GH actions, see #606 and #607.
|
process
|
release this is a coordinating issue for the release issues for this release are listed here the artifacts for the previous release were built manually for this release i would like to build and deploy artifacts using gh actions see and
| 1
|
365,346
| 10,780,931,830
|
IssuesEvent
|
2019-11-04 13:59:53
|
k8smeetup/website-tasks
|
https://api.github.com/repos/k8smeetup/website-tasks
|
closed
|
/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md
|
finished lang/zh priority/P0 sync/new version/1.16
|
Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md)
|
1.0
|
/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_reset_phase_preflight.md)
|
non_process
|
docs reference setup tools kubeadm generated kubeadm reset phase preflight md source file
| 0
|
14,369
| 17,394,251,102
|
IssuesEvent
|
2021-08-02 11:26:29
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Source Control Sync Job failed with "The property 'data' cannot be found on this object" message
|
Pri2 automation/svc cxp process-automation/subsvc product-issue triaged
|
We have a Source Control Sync setup with GitHub. Until recently it was working properly.
Even though the sync job is shown as completed, there is an error in the logs:
```
System.Management.Automation.PropertyNotFoundException: The property 'data' cannot be found on this object. Verify that the property exists.
at System.Management.Automation.ExceptionHandlingOps.CheckActionPreference(FunctionContext funcContext, Exception exception)
at System.Management.Automation.Interpreter.ActionCallInstruction`2.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.Interpreter.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.LightLambda.RunVoid1[T0](T0 arg0)
at System.Management.Automation.PSScriptCmdlet.RunClause(Action`1 clause, Object dollarUnderbar, Object inputToProcess)
at System.Management.Automation.PSScriptCmdlet.DoEndProcessing()
at System.Management.Automation.CommandProcessorBase.Complete()
```
The sync of course is not being done.
Thank you!
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
Source Control Sync Job failed with "The property 'data' cannot be found on this object" message - We have a Source Control Sync setup with GitHub. Until recently it was working properly.
Even though the sync job is shown as completed, there is an error in the logs:
```
System.Management.Automation.PropertyNotFoundException: The property 'data' cannot be found on this object. Verify that the property exists.
at System.Management.Automation.ExceptionHandlingOps.CheckActionPreference(FunctionContext funcContext, Exception exception)
at System.Management.Automation.Interpreter.ActionCallInstruction`2.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.EnterTryCatchFinallyInstruction.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.Interpreter.Run(InterpretedFrame frame)
at System.Management.Automation.Interpreter.LightLambda.RunVoid1[T0](T0 arg0)
at System.Management.Automation.PSScriptCmdlet.RunClause(Action`1 clause, Object dollarUnderbar, Object inputToProcess)
at System.Management.Automation.PSScriptCmdlet.DoEndProcessing()
at System.Management.Automation.CommandProcessorBase.Complete()
```
The sync of course is not being done.
Thank you!
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
source control sync job failed with the property data cannot be found on this object message we have a source control sync setup with github until recently it was working properly even though the sync job is shown as completed there is an error in the logs system management automation propertynotfoundexception the property data cannot be found on this object verify that the property exists at system management automation exceptionhandlingops checkactionpreference functioncontext funccontext exception exception at system management automation interpreter actioncallinstruction run interpretedframe frame at system management automation interpreter entertrycatchfinallyinstruction run interpretedframe frame at system management automation interpreter entertrycatchfinallyinstruction run interpretedframe frame at system management automation interpreter interpreter run interpretedframe frame at system management automation interpreter lightlambda at system management automation psscriptcmdlet runclause action clause object dollarunderbar object inputtoprocess at system management automation psscriptcmdlet doendprocessing at system management automation commandprocessorbase complete the sync of course is not being done thank you document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
20,005
| 26,479,558,733
|
IssuesEvent
|
2023-01-17 13:46:25
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
bazel-1.0.0: Depends: python but it is not installable
|
P2 type: support / not a bug (process) team-OSS
|
### Description of the bug:
I'm trying the install bazel 1.0.0 (a C++ package depends on this version specifically). When I attempt to install it I get the following message:
`$ sudo apt install bazel-1.0.0
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
Some packages could not be installed. This may mean that you have
requested an impossible situation or if you are using the unstable
distribution that some required packages have not yet been created
or been moved out of Incoming.
The following information may help to resolve the situation:
The following packages have unmet dependencies:
bazel-1.0.0 : Depends: python but it is not installable
E: Unable to correct problems, you have held broken packages.`
I have python 3.10.6 installed on my system:
`python
Python 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> exit()`
All other dependences are installed. My system is Ubuntu command line installed through WSL2:
`$ lsb_release -a
No LSB modules are available.
Distributor ID: Ubuntu
Description: Ubuntu 22.04.1 LTS
Release: 22.04
Codename: jammy`
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Attempt to install Bazel 1.0.0 on Ubuntu 22.04 through WSL2 on Windows 11.
### Which operating system are you running Bazel on?
Ubuntu 22.04 - jammy
### What is the output of `bazel info release`?
NA
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
NA
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
```text
NA
```
### Have you found anything relevant by searching the web?
There was a closed submission on this site. No resolution was listed.
### Any other information, logs, or outputs that you want to share?
_No response_
|
1.0
|
bazel-1.0.0: Depends: python but it is not installable - ### Description of the bug:
I'm trying the install bazel 1.0.0 (a C++ package depends on this version specifically). When I attempt to install it I get the following message:
`$ sudo apt install bazel-1.0.0
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
Some packages could not be installed. This may mean that you have
requested an impossible situation or if you are using the unstable
distribution that some required packages have not yet been created
or been moved out of Incoming.
The following information may help to resolve the situation:
The following packages have unmet dependencies:
bazel-1.0.0 : Depends: python but it is not installable
E: Unable to correct problems, you have held broken packages.`
I have python 3.10.6 installed on my system:
`python
Python 3.10.6 (main, Nov 14 2022, 16:10:14) [GCC 11.3.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> exit()`
All other dependences are installed. My system is Ubuntu command line installed through WSL2:
`$ lsb_release -a
No LSB modules are available.
Distributor ID: Ubuntu
Description: Ubuntu 22.04.1 LTS
Release: 22.04
Codename: jammy`
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Attempt to install Bazel 1.0.0 on Ubuntu 22.04 through WSL2 on Windows 11.
### Which operating system are you running Bazel on?
Ubuntu 22.04 - jammy
### What is the output of `bazel info release`?
NA
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
NA
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
```text
NA
```
### Have you found anything relevant by searching the web?
There was a closed submission on this site. No resolution was listed.
### Any other information, logs, or outputs that you want to share?
_No response_
|
process
|
bazel depends python but it is not installable description of the bug i m trying the install bazel a c package depends on this version specifically when i attempt to install it i get the following message sudo apt install bazel reading package lists done building dependency tree done reading state information done some packages could not be installed this may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of incoming the following information may help to resolve the situation the following packages have unmet dependencies bazel depends python but it is not installable e unable to correct problems you have held broken packages i have python installed on my system python python main nov on linux type help copyright credits or license for more information exit all other dependences are installed my system is ubuntu command line installed through lsb release a no lsb modules are available distributor id ubuntu description ubuntu lts release codename jammy what s the simplest easiest way to reproduce this bug please provide a minimal example if possible attempt to install bazel on ubuntu through on windows which operating system are you running bazel on ubuntu jammy what is the output of bazel info release na if bazel info release returns development version or non git tell us how you built bazel na what s the output of git remote get url origin git rev parse master git rev parse head text na have you found anything relevant by searching the web there was a closed submission on this site no resolution was listed any other information logs or outputs that you want to share no response
| 1
|
345,713
| 24,872,356,160
|
IssuesEvent
|
2022-10-27 16:07:45
|
Telefonica/mistica-design
|
https://api.github.com/repos/Telefonica/mistica-design
|
opened
|
toggleButton component
|
component: buttons library: mobile library: desktop category: documentation platform: ios platform: android platform: web category: materials new
|
- [ ] Specs
- [ ] Design documentation
- [ ] Add to mobile library
- [ ] Add to desktop library
- [ ] Add to UI Kit
- [ ] Android ticket
- [ ] iOS ticket
- [ ] Web ticket
|
1.0
|
toggleButton component - - [ ] Specs
- [ ] Design documentation
- [ ] Add to mobile library
- [ ] Add to desktop library
- [ ] Add to UI Kit
- [ ] Android ticket
- [ ] iOS ticket
- [ ] Web ticket
|
non_process
|
togglebutton component specs design documentation add to mobile library add to desktop library add to ui kit android ticket ios ticket web ticket
| 0
|
117,774
| 25,194,327,030
|
IssuesEvent
|
2022-11-12 10:00:10
|
rocky/python-decompile3
|
https://api.github.com/repos/rocky/python-decompile3
|
closed
|
Parse error at or near `DUP_TOP' (python 3.8)
|
invalid bytecode suspicious activity
|
When I decompiled the pyc, got the following error messages.
OS : win 10
Python 3.8.6
uncompyle6 : 3.7.4
decompyle3 : 3.3.2
pydisasm version 5.0.4
Python bytecode 3.8 (3413)
Source code size mod 2**32
```
Instruction context:
L. 5 52 LOAD_NAME list
54 LOAD_GENEXPR '<code_object <genexpr>>'
56 LOAD_STR '<genexpr>'
58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
60 LOAD_GENEXPR '<code_object <genexpr>>'
62 LOAD_STR '<genexpr>'
64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
66 LOAD_NAME f_in
68 GET_ITER
70 CALL_FUNCTION_1 1 ''
72 GET_ITER
74 CALL_FUNCTION_1 1 ''
76 CALL_FUNCTION_1 1 ''
78 STORE_NAME requirements
80 POP_BLOCK
82 LOAD_CONST None
-> 84 DUP_TOP
86 DUP_TOP
88 CALL_FUNCTION_3 3 ''
90 POP_TOP
92 JUMP_FORWARD 110 'to 110'
94_0 COME_FROM_WITH 48 '48'
94 <49>
96 POP_JUMP_IF_TRUE 100 'to 100'
98 <48>
100_0 COME_FROM 96 '96'
100 POP_TOP
102 POP_TOP
104 POP_TOP
106 POP_EXCEPT
108 POP_TOP
110_0 COME_FROM 92 '92'
# file start.pyc
# --- This code section failed: ---
L. 2 0 LOAD_CONST 0
2 LOAD_CONST None
4 IMPORT_NAME re
6 STORE_NAME re
8 LOAD_CONST 0
10 LOAD_CONST None
12 IMPORT_NAME os
14 STORE_NAME os
16 LOAD_CONST 0
18 LOAD_CONST None
20 IMPORT_NAME sys
22 STORE_NAME sys
24 LOAD_CONST 0
26 LOAD_CONST None
28 IMPORT_NAME pkg_resources
30 STORE_NAME pkg_resources
32 LOAD_CONST 0
34 LOAD_CONST None
36 IMPORT_NAME random
38 STORE_NAME random
L. 3 40 SETUP_FINALLY 228 'to 228'
L. 4 42 LOAD_NAME open
44 LOAD_STR 'requirements.txt'
46 CALL_FUNCTION_1 1 ''
48 SETUP_WITH 94 'to 94'
50 STORE_NAME f_in
L. 5 52 LOAD_NAME list
54 LOAD_GENEXPR '<code_object <genexpr>>'
56 LOAD_STR '<genexpr>'
58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
60 LOAD_GENEXPR '<code_object <genexpr>>'
62 LOAD_STR '<genexpr>'
64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
66 LOAD_NAME f_in
68 GET_ITER
70 CALL_FUNCTION_1 1 ''
72 GET_ITER
74 CALL_FUNCTION_1 1 ''
76 CALL_FUNCTION_1 1 ''
78 STORE_NAME requirements
80 POP_BLOCK
82 LOAD_CONST None
84 DUP_TOP
86 DUP_TOP
88 CALL_FUNCTION_3 3 ''
90 POP_TOP
92 JUMP_FORWARD 110 'to 110'
94_0 COME_FROM_WITH 48 '48'
94 <49>
96 POP_JUMP_IF_TRUE 100 'to 100'
98 <48>
100_0 COME_FROM 96 '96'
100 POP_TOP
102 POP_TOP
104 POP_TOP
106 POP_EXCEPT
108 POP_TOP
110_0 COME_FROM 92 '92'
L. 6 110 LOAD_NAME requirements
112 GET_ITER
114_0 COME_FROM 222 '222'
114_1 COME_FROM 210 '210'
114_2 COME_FROM 132 '132'
114 FOR_ITER 224 'to 224'
116 STORE_NAME require
L. 7 118 SETUP_FINALLY 134 'to 134'
L. 8 120 LOAD_NAME pkg_resources
122 LOAD_METHOD require
124 LOAD_NAME require
126 CALL_METHOD_1 1 ''
128 POP_TOP
130 POP_BLOCK
132 JUMP_BACK 114 'to 114'
134_0 COME_FROM_FINALLY 118 '118'
L. 9 134 DUP_TOP
136 LOAD_NAME Exception
138 <121> 220 ''
140 POP_TOP
142 STORE_NAME e
144 POP_TOP
146 SETUP_FINALLY 212 'to 212'
L. 10 148 LOAD_NAME re
150 LOAD_METHOD search
152 LOAD_STR "\\'(.*?)\\'"
154 LOAD_NAME str
156 LOAD_NAME e
158 CALL_FUNCTION_1 1 ''
160 CALL_METHOD_2 2 ''
162 LOAD_METHOD group
164 LOAD_CONST 1
166 CALL_METHOD_1 1 ''
168 STORE_NAME a
L. 11 170 LOAD_NAME print
172 LOAD_STR 'Installing '
174 LOAD_NAME a
176 BINARY_ADD
178 LOAD_STR ' ...'
180 BINARY_ADD
182 CALL_FUNCTION_1 1 ''
184 POP_TOP
L. 12 186 LOAD_NAME os
188 LOAD_METHOD system
190 LOAD_STR 'pip install '
192 LOAD_NAME a
194 BINARY_ADD
196 CALL_METHOD_1 1 ''
198 POP_TOP
200 POP_BLOCK
202 POP_EXCEPT
204 LOAD_CONST None
206 STORE_NAME e
208 DELETE_NAME e
210 JUMP_BACK 114 'to 114'
212_0 COME_FROM_FINALLY 146 '146'
212 LOAD_CONST None
214 STORE_NAME e
216 DELETE_NAME e
218 <48>
220 <48>
222 JUMP_BACK 114 'to 114'
224_0 COME_FROM 114 '114'
224 POP_BLOCK
226 JUMP_FORWARD 280 'to 280'
228_0 COME_FROM_FINALLY 40 '40'
L. 13 228 DUP_TOP
230 LOAD_NAME Exception
232_234 <121> 278 ''
236 POP_TOP
238 STORE_NAME e
240 POP_TOP
242 SETUP_FINALLY 270 'to 270'
L. 14 244 LOAD_NAME print
246 LOAD_STR 'installation error: '
248 LOAD_NAME e
250 FORMAT_VALUE 0 ''
252 BUILD_STRING_2 2
254 CALL_FUNCTION_1 1 ''
256 POP_TOP
258 POP_BLOCK
260 POP_EXCEPT
262 LOAD_CONST None
264 STORE_NAME e
266 DELETE_NAME e
268 JUMP_FORWARD 280 'to 280'
270_0 COME_FROM_FINALLY 242 '242'
270 LOAD_CONST None
272 STORE_NAME e
274 DELETE_NAME e
276 <48>
278 <48>
280_0 COME_FROM 268 '268'
280_1 COME_FROM 226 '226'
L. 16 280 LOAD_CONST 0
282 LOAD_CONST None
284 IMPORT_NAME asyncio
286 STORE_NAME asyncio
288 LOAD_CONST 0
290 LOAD_CONST None
292 IMPORT_NAME logging
294 STORE_NAME logging
296 LOAD_CONST 0
298 LOAD_CONST None
300 IMPORT_NAME requests
302 STORE_NAME requests
304 LOAD_CONST 0
306 LOAD_CONST None
308 IMPORT_NAME sqlite3
310 STORE_NAME sqlite3
312 LOAD_CONST 0
314 LOAD_CONST None
316 IMPORT_NAME base64
318 STORE_NAME base64
L. 17 320 LOAD_CONST 0
322 LOAD_CONST ('sleep', 'strptime')
324 IMPORT_NAME time
326 IMPORT_FROM sleep
328 STORE_NAME sleep
330 IMPORT_FROM strptime
332 STORE_NAME strptime
334 POP_TOP
L. 18 336 LOAD_CONST 0
338 LOAD_CONST ('KeyboardButtonUrl', 'MessageMediaContact', 'ReplyInlineMarkup', 'UpdateShortMessage')
340 IMPORT_NAME_ATTR telethon.tl.types
342 IMPORT_FROM KeyboardButtonUrl
344 STORE_NAME KeyboardButtonUrl
346 IMPORT_FROM MessageMediaContact
348 STORE_NAME MessageMediaContact
350 IMPORT_FROM ReplyInlineMarkup
352 STORE_NAME ReplyInlineMarkup
354 IMPORT_FROM UpdateShortMessage
356 STORE_NAME UpdateShortMessage
358 POP_TOP
L. 19 360 LOAD_CONST 0
362 LOAD_CONST ('TelegramClient', 'client', 'errors', 'events', 'functions', 'connection')
364 IMPORT_NAME telethon
366 IMPORT_FROM TelegramClient
368 STORE_NAME TelegramClient
370 IMPORT_FROM client
372 STORE_GLOBAL client
374 IMPORT_FROM errors
376 STORE_NAME errors
378 IMPORT_FROM events
380 STORE_NAME events
382 IMPORT_FROM functions
384 STORE_NAME functions
386 IMPORT_FROM connection
388 STORE_NAME connection
390 POP_TOP
L. 20 392 LOAD_CONST 0
394 LOAD_CONST ('JoinChannelRequest',)
396 IMPORT_NAME_ATTR telethon.tl.functions.channels
398 IMPORT_FROM JoinChannelRequest
400 STORE_NAME JoinChannelRequest
402 POP_TOP
L. 21 404 LOAD_CONST 0
406 LOAD_CONST ('GetBotCallbackAnswerRequest', 'GetHistoryRequest', 'ImportChatInviteRequest', 'GetMessagesViewsRequest')
408 IMPORT_NAME_ATTR telethon.tl.functions.messages
410 IMPORT_FROM GetBotCallbackAnswerRequest
412 STORE_NAME GetBotCallbackAnswerRequest
414 IMPORT_FROM GetHistoryRequest
416 STORE_NAME GetHistoryRequest
418 IMPORT_FROM ImportChatInviteRequest
420 STORE_NAME ImportChatInviteRequest
422 IMPORT_FROM GetMessagesViewsRequest
424 STORE_NAME GetMessagesViewsRequest
426 POP_TOP
L. 22 428 LOAD_CONST 0
430 LOAD_CONST ('datetime', 'timedelta')
432 IMPORT_NAME datetime
434 IMPORT_FROM datetime
436 STORE_NAME datetime
438 IMPORT_FROM timedelta
440 STORE_NAME timedelta
442 POP_TOP
L. 23 444 LOAD_CONST 0
446 LOAD_CONST ('Fore', 'Style', 'init')
448 IMPORT_NAME colorama
450 IMPORT_FROM Fore
452 STORE_NAME Fore
454 IMPORT_FROM Style
456 STORE_NAME Style
458 IMPORT_FROM init
460 STORE_NAME color_ama
462 POP_TOP
L. 24 464 LOAD_CONST 0
466 LOAD_CONST ('BeautifulSoup',)
468 IMPORT_NAME bs4
470 IMPORT_FROM BeautifulSoup
472 STORE_NAME BeautifulSoup
474 POP_TOP
L. 25 476 LOAD_CONST 0
478 LOAD_CONST None
480 IMPORT_NAME random
482 STORE_NAME random
L. 26 484 LOAD_CONST 0
486 LOAD_CONST None
488 IMPORT_NAME socks
490 STORE_NAME socks
L. 27 492 LOAD_NAME logging
494 LOAD_ATTR basicConfig
496 LOAD_NAME logging
498 LOAD_ATTR ERROR
500 LOAD_CONST ('level',)
502 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args'
504 POP_TOP
L. 28 506 LOAD_NAME color_ama
508 LOAD_CONST True
510 LOAD_CONST ('autoreset',)
512 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args'
514 POP_TOP
L. 29 516 LOAD_NAME os
518 LOAD_METHOD system
520 LOAD_NAME os
522 LOAD_ATTR name
524 LOAD_STR 'nt'
526 COMPARE_OP ==
528_530 POP_JUMP_IF_FALSE 536 'to 536'
532 LOAD_STR 'cls'
534 JUMP_FORWARD 538 'to 538'
536_0 COME_FROM 528 '528'
536 LOAD_STR 'clear'
538_0 COME_FROM 534 '534'
538 CALL_METHOD_1 1 ''
540 POP_TOP
L. 30 542 LOAD_STR '\n'
L. 31 544 LOAD_NAME Style
546 LOAD_ATTR NORMAL
548 LOAD_NAME Fore
550 LOAD_ATTR MAGENTA
552 BINARY_ADD
L. 30 554 FORMAT_VALUE 0 ''
556 LOAD_STR ' ____ ___ ___ _ _ '
L. 31 558 LOAD_NAME Fore
560 LOAD_ATTR GREEN
L. 30 562 FORMAT_VALUE 0 ''
564 LOAD_STR ' ___ ___ \n'
L. 31 566 LOAD_NAME Style
568 LOAD_ATTR NORMAL
570 LOAD_NAME Fore
572 LOAD_ATTR MAGENTA
574 BINARY_ADD
L. 30 576 FORMAT_VALUE 0 ''
578 LOAD_STR ' / ___| / _ \\ |_ _| | \\ | | '
L. 31 580 LOAD_NAME Fore
582 LOAD_ATTR GREEN
L. 30 584 FORMAT_VALUE 0 ''
586 LOAD_STR ' / _ \\ ( _ ) \n'
L. 31 588 LOAD_NAME Style
590 LOAD_ATTR NORMAL
592 LOAD_NAME Fore
594 LOAD_ATTR MAGENTA
596 BINARY_ADD
L. 30 598 FORMAT_VALUE 0 ''
600 LOAD_STR ' | | | | | | | | | \\| | '
L. 31 602 LOAD_NAME Fore
604 LOAD_ATTR GREEN
L. 30 606 FORMAT_VALUE 0 ''
608 LOAD_STR '| (_) | / _ \\ \n'
L. 31 610 LOAD_NAME Style
612 LOAD_ATTR NORMAL
614 LOAD_NAME Fore
616 LOAD_ATTR MAGENTA
618 BINARY_ADD
L. 30 620 FORMAT_VALUE 0 ''
622 LOAD_STR ' | |___ | |_| | | | | |\\ | '
L. 31 624 LOAD_NAME Fore
626 LOAD_ATTR GREEN
L. 30 628 FORMAT_VALUE 0 ''
630 LOAD_STR ' \\__, | | (_) |\n'
L. 31 632 LOAD_NAME Style
634 LOAD_ATTR NORMAL
636 LOAD_NAME Fore
638 LOAD_ATTR MAGENTA
640 BINARY_ADD
L. 30 642 FORMAT_VALUE 0 ''
644 LOAD_STR ' \\____| \\___/ |___| |_| \\_| '
L. 31 646 LOAD_NAME Fore
648 LOAD_ATTR GREEN
L. 30 650 FORMAT_VALUE 0 ''
652 LOAD_STR ' /_/ \\___/ \n'
L. 36 654 LOAD_NAME Fore
656 LOAD_ATTR BLUE
L. 30 658 FORMAT_VALUE 0 ''
660 LOAD_STR ' Edit By'
L. 36 662 LOAD_NAME Style
664 LOAD_ATTR DIM
666 LOAD_NAME Fore
668 LOAD_ATTR RED
670 BINARY_ADD
L. 30 672 FORMAT_VALUE 0 ''
674 LOAD_STR ':'
L. 36 676 LOAD_NAME Fore
678 LOAD_ATTR YELLOW
L. 30 680 FORMAT_VALUE 0 ''
682 LOAD_STR ' Abbas Bachari '
L. 36 684 LOAD_NAME Fore
686 LOAD_ATTR BLUE
L. 30 688 FORMAT_VALUE 0 ''
690 LOAD_STR 'Version'
L. 36 692 LOAD_NAME Fore
694 LOAD_ATTR YELLOW
L. 30 696 FORMAT_VALUE 0 ''
698 LOAD_STR ' 7.9\n'
L. 37 700 LOAD_NAME Style
702 LOAD_ATTR NORMAL
704 LOAD_NAME Fore
706 LOAD_ATTR RED
708 BINARY_ADD
L. 30 710 FORMAT_VALUE 0 ''
712 LOAD_STR '==============================================\n'
L. 38 714 LOAD_NAME Style
716 LOAD_ATTR BRIGHT
718 LOAD_NAME Fore
720 LOAD_ATTR GREEN
722 BINARY_ADD
L. 30 724 FORMAT_VALUE 0 ''
726 LOAD_STR 'Sponsor Channel '
L. 36 728 LOAD_NAME Style
730 LOAD_ATTR DIM
732 LOAD_NAME Fore
734 LOAD_ATTR RED
736 BINARY_ADD
L. 30 738 FORMAT_VALUE 0 ''
740 LOAD_STR ':'
L. 38 742 LOAD_NAME Style
744 LOAD_ATTR RESET_ALL
L. 30 746 FORMAT_VALUE 0 ''
748 LOAD_STR ' @COIN98'
750 BUILD_STRING_39 39
752 STORE_NAME banner
L. 41 754 LOAD_CONST 799906641
756 LOAD_STR 'BitcoinClick_bot'
758 LOAD_STR '/start 2fv0'
760 LOAD_CONST ('id', 'username', 'start')
762 BUILD_CONST_KEY_MAP_3 3
L. 42 764 LOAD_CONST 741849360
766 LOAD_STR 'Litecoin_click_bot'
768 LOAD_STR '/start aZYG'
770 LOAD_CONST ('id', 'username', 'start')
772 BUILD_CONST_KEY_MAP_3 3
L. 43 774 LOAD_CONST 715510199
776 LOAD_STR 'Dogecoin_click_bot'
778 LOAD_STR '/start ljSP'
780 LOAD_CONST ('id', 'username', 'start')
782 BUILD_CONST_KEY_MAP_3 3
L. 44 784 LOAD_CONST 687127269
786 LOAD_STR 'BCH_clickbot'
788 LOAD_STR '/start 3su3'
790 LOAD_CONST ('id', 'username', 'start')
792 BUILD_CONST_KEY_MAP_3 3
L. 45 794 LOAD_CONST 850081470
796 LOAD_STR 'Zcash_click_bot'
798 LOAD_STR '/start 6qKG'
800 LOAD_CONST ('id', 'username', 'start')
802 BUILD_CONST_KEY_MAP_3 3
L. 40 804 LOAD_CONST ('BTC', 'LTC', 'DOGE', 'BCH', 'ZEC')
806 BUILD_CONST_KEY_MAP_5 5
808 STORE_NAME Symbols
L. 48 810 LOAD_CONST 197597
812 STORE_GLOBAL api_id
L. 49 814 LOAD_STR '227647fae1d2a3a1419bdd527337c87d'
816 STORE_GLOBAL api_hash
L. 50 818 LOAD_STR 'RGVhciB1c2VyLCB5b3UgbmVlZCB0byBzdWJzY3JpYmUgdG8gdGhlCiAgICBmb2xsb3dpbmcgVGVsZWdyYW0gY2hhbm5lbBtbMTswbQogICAgQ2hhbm5lbCBJRCA6IEBDT0lOOTgNCgoK'
820 STORE_NAME MSG
L. 51 822 LOAD_NAME requests
824 LOAD_METHOD session
826 CALL_METHOD_0 0 ''
828 STORE_GLOBAL session
L. 52 830 LOAD_STR 'User-Agent'
832 LOAD_STR 'Mozilla/5.0 (Linux; Android 9; moto g(7) play) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.111 Mobile Safari/537.36'
834 BUILD_MAP_1 1
836 STORE_NAME ua
L. 53 838 LOAD_STR ''
840 STORE_GLOBAL dogeclick_channel
L. 54 842 LOAD_STR ''
844 STORE_GLOBAL phone_number
L. 55 846 LOAD_STR ''
848 STORE_GLOBAL Wallet
L. 56 850 LOAD_STR ''
852 STORE_GLOBAL isVisit
L. 57 854 LOAD_STR ''
856 STORE_GLOBAL isJoin
L. 58 858 LOAD_STR ''
860 STORE_GLOBAL isBotMs
L. 59 862 LOAD_STR ''
864 STORE_GLOBAL isLoop
L. 60 866 LOAD_STR ''
868 STORE_GLOBAL Balance
L. 61 870 BUILD_MAP_0 0
872 STORE_GLOBAL SET
L. 62 874 LOAD_STR ''
876 STORE_GLOBAL coin
L. 63 878 LOAD_CONST 0
880 STORE_NAME INDEX
L. 64 882 LOAD_CONST ()
884 STORE_GLOBAL NUM
L. 65 886 LOAD_STR 'SIT'
888 STORE_GLOBAL stat
L. 66 890 LOAD_CONST None
892 STORE_GLOBAL captcha
L. 67 894 LOAD_CONST None
896 STORE_GLOBAL sit
L. 68 898 LOAD_STR ''
900 STORE_GLOBAL message
L. 69 902 LOAD_STR ''
904 STORE_GLOBAL channel_name
L. 70 906 LOAD_CONST None
908 STORE_GLOBAL channel_id
L. 71 910 LOAD_STR ''
912 STORE_GLOBAL bot_name
L. 72 914 LOAD_STR ''
916 STORE_GLOBAL COMENT
L. 73 918 LOAD_CONST 7.85
920 STORE_NAME APVER
L. 74 922 BUILD_LIST_0 0
924 STORE_GLOBAL BOTS
L. 75 926 BUILD_LIST_0 0
928 STORE_GLOBAL CHNS
L. 76 930 BUILD_LIST_0 0
932 STORE_GLOBAL GRPS
L. 77 934 BUILD_LIST_0 0
936 STORE_GLOBAL USRS
L. 78 938 LOAD_CONST 0
940 STORE_GLOBAL LIMIT
L. 79 942 LOAD_STR ''
944 STORE_GLOBAL LICE
L. 80 946 LOAD_STR ''
948 STORE_GLOBAL MASE
L. 81 950 LOAD_CONST 0
952 STORE_GLOBAL VER
L. 82 954 LOAD_CONST 0
956 STORE_GLOBAL mass_id
L. 83 958 LOAD_CONST None
960 STORE_GLOBAL click_data
L. 84 962 LOAD_STR ''
964 STORE_GLOBAL VIPMES
L. 85 966 LOAD_CONST 0
968 STORE_GLOBAL JNUM
L. 86 970 LOAD_CONST 0
972 STORE_GLOBAL ERN
L. 87 974 LOAD_CONST 0
976 STORE_GLOBAL WIWNUM
L. 88 978 LOAD_CONST 0
980 STORE_GLOBAL WIWCON
L. 89 982 LOAD_CONST 0
984 STORE_NAME LOPS
L. 90 986 BUILD_MAP_0 0
988 STORE_GLOBAL SETIN
L. 91 990 BUILD_LIST_0 0
992 STORE_GLOBAL PLOGINS
L. 92 994 BUILD_MAP_0 0
996 STORE_GLOBAL MESEGS
L. 93 998 BUILD_MAP_0 0
1000 STORE_GLOBAL ADMESEG
L. 94 1002 LOAD_STR 'YES'
1004 STORE_GLOBAL ADD
L. 95 1006 LOAD_CONST None
1008 STORE_GLOBAL MINWD
L. 96 1010 LOAD_STR 'NUMBERS.txt'
1012 STORE_GLOBAL path
L. 97 1014 LOAD_CONST None
1016 STORE_GLOBAL Proxy
L. 99 1018 LOAD_CODE <code_object isNumber>
1020 LOAD_STR 'isNumber'
1022 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1024 STORE_NAME isNumber
L. 105 1026 LOAD_CODE <code_object request>
1028 LOAD_STR 'request'
1030 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1032 STORE_NAME request
L. 131 1034 LOAD_CODE <code_object GetIP>
1036 LOAD_STR 'GetIP'
1038 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1040 STORE_NAME GetIP
L. 145 1042 LOAD_CODE <code_object SEND_CONTACT>
1044 LOAD_STR 'SEND_CONTACT'
1046 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1048 STORE_NAME SEND_CONTACT
L. 159 1050 LOAD_CODE <code_object send_messege_to_members>
1052 LOAD_STR 'send_messege_to_members'
1054 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1056 STORE_NAME send_messege_to_members
L. 203 1058 LOAD_CODE <code_object CHNUM>
1060 LOAD_STR 'CHNUM'
1062 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1064 STORE_NAME CHNUM
L. 211 1066 LOAD_CODE <code_object numbers>
1068 LOAD_STR 'numbers'
1070 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1072 STORE_NAME numbers
L. 272 1074 LOAD_CODE <code_object AnonsRobot>
1076 LOAD_STR 'AnonsRobot'
1078 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1080 STORE_NAME AnonsRobot
L. 285 1082 LOAD_CODE <code_object START_PLOGINS>
1084 LOAD_STR 'START_PLOGINS'
1086 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1088 STORE_NAME START_PLOGINS
L. 382 1090 LOAD_CODE <code_object SEND_PLOGIN_MESSEGE>
1092 LOAD_STR 'SEND_PLOGIN_MESSEGE'
1094 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1096 STORE_NAME SEND_PLOGIN_MESSEGE
L. 403 1098 LOAD_CODE <code_object SEND_GROUP_MESSEGE>
1100 LOAD_STR 'SEND_GROUP_MESSEGE'
1102 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1104 STORE_NAME SEND_GROUP_MESSEGE
L. 435 1106 LOAD_CODE <code_object Settings>
1108 LOAD_STR 'Settings'
1110 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1112 STORE_NAME Settings
L. 569 1114 LOAD_CODE <code_object SEND_TO_PRIVAIT_GROUP>
1116 LOAD_STR 'SEND_TO_PRIVAIT_GROUP'
1118 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1120 STORE_NAME SEND_TO_PRIVAIT_GROUP
L. 611 1122 LOAD_CODE <code_object Creat_user>
1124 LOAD_STR 'Creat_user'
1126 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1128 STORE_NAME Creat_user
L. 617 1130 LOAD_CODE <code_object Add_Plogin>
1132 LOAD_STR 'Add_Plogin'
1134 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1136 STORE_NAME Add_Plogin
L. 629 1138 LOAD_CODE <code_object Delete_Plogin>
1140 LOAD_STR 'Delete_Plogin'
1142 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1144 STORE_NAME Delete_Plogin
L. 635 1146 LOAD_CODE <code_object GetLoop>
1148 LOAD_STR 'GetLoop'
1150 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1152 STORE_NAME GetLoop
L. 648 1154 LOAD_CODE <code_object Add_Channel>
1156 LOAD_STR 'Add_Channel'
1158 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1160 STORE_NAME Add_Channel
L. 663 1162 LOAD_STR 'NO'
1164 STORE_GLOBAL ISWID
L. 664 1166 LOAD_CODE <code_object Delete_Cannels>
1168 LOAD_STR 'Delete_Cannels'
1170 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1172 STORE_NAME Delete_Cannels
L. 711 1174 LOAD_CODE <code_object reset>
1176 LOAD_STR 'reset'
1178 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1180 STORE_NAME reset
L. 736 1182 LOAD_CODE <code_object maseg>
1184 LOAD_STR 'maseg'
1186 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1188 STORE_NAME maseg
L. 740 1190 LOAD_CODE <code_object WAIT>
1192 LOAD_STR 'WAIT'
1194 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1196 STORE_NAME WAIT
L. 777 1198 LOAD_CODE <code_object FloodTimer>
1200 LOAD_STR 'FloodTimer'
1202 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1204 STORE_NAME FloodTimer
L. 787 1206 LOAD_CODE <code_object sleepTime>
1208 LOAD_STR 'sleepTime'
1210 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1212 STORE_NAME sleepTime
L. 806 1214 LOAD_CODE <code_object JoinTimer>
1216 LOAD_STR 'JoinTimer'
1218 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1220 STORE_NAME JoinTimer
L. 816 1222 LOAD_CODE <code_object GET_USERNAME>
1224 LOAD_STR 'GET_USERNAME'
1226 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1228 STORE_NAME GET_USERNAME
L. 839 1230 LOAD_CODE <code_object forward_messages>
1232 LOAD_STR 'forward_messages'
1234 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1236 STORE_NAME forward_messages
L. 854 1238 LOAD_CODE <code_object start>
1240 LOAD_STR 'start'
1242 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1244 STORE_NAME start
L. 997 1246 LOAD_CODE <code_object main>
1248 LOAD_STR 'main'
1250 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1252 STORE_NAME main
1254_0 COME_FROM 1756 '1756'
1254_1 COME_FROM 1752 '1752'
1254_2 COME_FROM 1718 '1718'
L.1293 1254_1256 SETUP_FINALLY 1720 'to 1720'
L.1294 1258 LOAD_NAME asyncio
1260 LOAD_METHOD get_event_loop
1262 CALL_METHOD_0 0 ''
1264 LOAD_METHOD run_until_complete
1266 LOAD_NAME numbers
1268 CALL_FUNCTION_0 0 ''
1270 CALL_METHOD_1 1 ''
1272 POP_TOP
L.1295 1274 LOAD_NAME LOPS
1276 LOAD_CONST 1
1278 INPLACE_ADD
1280 STORE_NAME LOPS
L.1296 1282 LOAD_GLOBAL NUM
1284 GET_ITER
1286_0 COME_FROM 1328 '1328'
1286 FOR_ITER 1332 'to 1332'
1288 STORE_NAME i
L.1297 1290 LOAD_NAME INDEX
1292 LOAD_CONST 1
1294 INPLACE_ADD
1296 STORE_NAME INDEX
L.1298 1298 LOAD_CONST 0
1300 STORE_GLOBAL JNUM
L.1299 1302 LOAD_NAME asyncio
1304 LOAD_METHOD get_event_loop
1306 CALL_METHOD_0 0 ''
1308 LOAD_METHOD run_until_complete
1310 LOAD_NAME main
1312 LOAD_NAME i
1314 CALL_FUNCTION_1 1 ''
1316 CALL_METHOD_1 1 ''
1318 POP_TOP
L.1302 1320 LOAD_NAME sleep
1322 LOAD_CONST 3
1324 CALL_FUNCTION_1 1 ''
1326 POP_TOP
1328_1330 JUMP_BACK 1286 'to 1286'
1332_0 COME_FROM 1286 '1286'
L.1303 1332 LOAD_NAME os
1334 LOAD_METHOD system
1336 LOAD_NAME os
1338 LOAD_ATTR name
1340 LOAD_STR 'nt'
1342 COMPARE_OP ==
1344_1346 POP_JUMP_IF_FALSE 1352 'to 1352'
1348 LOAD_STR 'cls'
1350 JUMP_FORWARD 1354 'to 1354'
1352_0 COME_FROM 1344 '1344'
1352 LOAD_STR 'clear'
1354_0 COME_FROM 1350 '1350'
1354 CALL_METHOD_1 1 ''
1356 POP_TOP
L.1304 1358 LOAD_NAME format
1360 LOAD_GLOBAL ERN
1362 LOAD_STR '.8f'
1364 CALL_FUNCTION_2 2 ''
1366 STORE_NAME ER
L.1305 1368 LOAD_NAME format
1370 LOAD_GLOBAL WIWCON
1372 LOAD_STR '.8f'
1374 CALL_FUNCTION_2 2 ''
1376 STORE_NAME WC
L.1306 1378 LOAD_GLOBAL LICE
1380 LOAD_STR 'VIP'
1382 COMPARE_OP ==
1384_1386 POP_JUMP_IF_FALSE 1406 'to 1406'
L.1307 1388 LOAD_GLOBAL LICE
1390 FORMAT_VALUE 0 ''
1392 LOAD_STR '('
1394 LOAD_GLOBAL LIMIT
1396 FORMAT_VALUE 0 ''
1398 LOAD_STR ')'
1400 BUILD_STRING_4 4
1402 STORE_NAME VI
1404 JUMP_FORWARD 1428 'to 1428'
1406_0 COME_FROM 1384 '1384'
L.1309 1406 LOAD_NAME Fore
1408 LOAD_ATTR RED
1410 FORMAT_VALUE 0 ''
1412 LOAD_GLOBAL LICE
1414 FORMAT_VALUE 0 ''
1416 LOAD_STR '('
1418 LOAD_GLOBAL LIMIT
1420 FORMAT_VALUE 0 ''
1422 LOAD_STR ')'
1424 BUILD_STRING_5 5
1426 STORE_NAME VI
1428_0 COME_FROM 1404 '1404'
L.1311 1428 LOAD_NAME print
1430 LOAD_NAME banner
1432 CALL_FUNCTION_1 1 ''
1434 POP_TOP
L.1312 1436 LOAD_NAME print
1438 LOAD_NAME Fore
1440 LOAD_ATTR GREEN
1442 FORMAT_VALUE 0 ''
1444 LOAD_STR 'Script License '
1446 LOAD_NAME Fore
1448 LOAD_ATTR RED
1450 FORMAT_VALUE 0 ''
1452 LOAD_STR ':'
1454 LOAD_NAME Fore
1456 LOAD_ATTR RESET
1458 FORMAT_VALUE 0 ''
1460 LOAD_STR ' '
1462 LOAD_NAME VI
1464 FORMAT_VALUE 0 ''
1466 BUILD_STRING_7 7
1468 CALL_FUNCTION_1 1 ''
1470 POP_TOP
L.1313 1472 LOAD_NAME print
1474 LOAD_NAME Fore
1476 LOAD_ATTR GREEN
1478 FORMAT_VALUE 0 ''
1480 LOAD_STR 'Total Loops '
1482 LOAD_NAME Fore
1484 LOAD_ATTR RED
1486 FORMAT_VALUE 0 ''
1488 LOAD_STR ':'
1490 LOAD_NAME Fore
1492 LOAD_ATTR RESET
1494 FORMAT_VALUE 0 ''
1496 LOAD_STR ' '
1498 LOAD_NAME LOPS
1500 FORMAT_VALUE 0 ''
1502 BUILD_STRING_7 7
1504 CALL_FUNCTION_1 1 ''
1506 POP_TOP
L.1314 1508 LOAD_NAME print
1510 LOAD_NAME Fore
1512 LOAD_ATTR GREEN
1514 FORMAT_VALUE 0 ''
1516 LOAD_STR 'Total accounts '
1518 LOAD_NAME Fore
1520 LOAD_ATTR RED
1522 FORMAT_VALUE 0 ''
1524 LOAD_STR ':'
1526 LOAD_NAME Fore
1528 LOAD_ATTR RESET
1530 FORMAT_VALUE 0 ''
1532 LOAD_STR ' '
1534 LOAD_NAME len
1536 LOAD_GLOBAL NUM
1538 CALL_FUNCTION_1 1 ''
1540 FORMAT_VALUE 0 ''
1542 BUILD_STRING_7 7
1544 CALL_FUNCTION_1 1 ''
1546 POP_TOP
L.1315 1548 LOAD_NAME print
1550 LOAD_NAME Fore
1552 LOAD_ATTR GREEN
1554 FORMAT_VALUE 0 ''
1556 LOAD_STR 'accounts withdraw '
1558 LOAD_NAME Fore
1560 LOAD_ATTR RED
1562 FORMAT_VALUE 0 ''
1564 LOAD_STR ':'
1566 LOAD_NAME Fore
1568 LOAD_ATTR RESET
1570 FORMAT_VALUE 0 ''
1572 LOAD_STR ' '
1574 LOAD_GLOBAL WIWNUM
1576 FORMAT_VALUE 0 ''
1578 LOAD_STR ' account'
1580 BUILD_STRING_8 8
1582 CALL_FUNCTION_1 1 ''
1584 POP_TOP
L.1316 1586 LOAD_NAME print
1588 LOAD_NAME Fore
1590 LOAD_ATTR GREEN
1592 FORMAT_VALUE 0 ''
1594 LOAD_STR 'Total withdraw '
1596 LOAD_NAME Fore
1598 LOAD_ATTR RED
1600 FORMAT_VALUE 0 ''
1602 LOAD_STR ':'
1604 LOAD_NAME Fore
1606 LOAD_ATTR RESET
1608 FORMAT_VALUE 0 ''
1610 LOAD_STR ' '
1612 LOAD_NAME WC
1614 FORMAT_VALUE 0 ''
1616 LOAD_STR ' '
1618 LOAD_GLOBAL coin
1620 FORMAT_VALUE 0 ''
1622 BUILD_STRING_9 9
1624 CALL_FUNCTION_1 1 ''
1626 POP_TOP
L.1317 1628 LOAD_NAME print
1630 LOAD_NAME Fore
1632 LOAD_ATTR GREEN
1634 FORMAT_VALUE 0 ''
1636 LOAD_STR 'Total balance '
1638 LOAD_NAME Fore
1640 LOAD_ATTR RED
1642 FORMAT_VALUE 0 ''
1644 LOAD_STR ':'
1646 LOAD_NAME Fore
1648 LOAD_ATTR RESET
1650 FORMAT_VALUE 0 ''
1652 LOAD_STR ' '
1654 LOAD_NAME ER
1656 FORMAT_VALUE 0 ''
1658 LOAD_STR ' '
1660 LOAD_GLOBAL coin
1662 FORMAT_VALUE 0 ''
1664 BUILD_STRING_9 9
1666 CALL_FUNCTION_1 1 ''
1668 POP_TOP
L.1318 1670 LOAD_NAME print
1672 LOAD_STR '\n'
1674 CALL_FUNCTION_1 1 ''
1676 POP_TOP
L.1319 1678 LOAD_GLOBAL SET
1680 LOAD_STR 'AUTOLOOP'
1682 BINARY_SUBSCR
1684 LOAD_METHOD upper
1686 CALL_METHOD_0 0 ''
1688 LOAD_STR 'YES'
1690 COMPARE_OP !=
1692_1694 POP_JUMP_IF_FALSE 1702 'to 1702'
L.1320 1696 POP_BLOCK
1698_1700 BREAK_LOOP 1760 'to 1760'
1702_0 COME_FROM 1692 '1692'
L.1322 1702 LOAD_NAME sleepTime
1704 CALL_FUNCTION_0 0 ''
1706 POP_TOP
L.1323 1708 LOAD_CONST 0
1710 STORE_GLOBAL ERN
L.1324 1712 LOAD_CONST 0
1714 STORE_NAME INDEX
1716 POP_BLOCK
1718 JUMP_BACK 1254 'to 1254'
1720_0 COME_FROM_FINALLY 1254 '1254'
L.1326 1720 DUP_TOP
1722 LOAD_NAME KeyboardInterrupt
1724_1726 <121> 1754 ''
1728 POP_TOP
1730 POP_TOP
1732 POP_TOP
L.1327 1734 LOAD_NAME print
1736 LOAD_STR ''
1738 CALL_FUNCTION_1 1 ''
1740 POP_TOP
L.1328 1742 LOAD_NAME sys
1744 LOAD_METHOD exit
1746 CALL_METHOD_0 0 ''
1748 POP_TOP
1750 POP_EXCEPT
1752 JUMP_BACK 1254 'to 1254'
1754 <48>
1756_1758 JUMP_BACK 1254 'to 1254'
1760_0 COME_FROM 1698 '1698'
Parse error at or near `DUP_TOP' instruction at offset 84
```
i also attached : .`pyc `file , decompiled results and `pydisasm `results
[decompiled_output.zip](https://github.com/rocky/python-decompile3/files/5471265/decompiled_output.zip)
[pycfile.zip](https://github.com/rocky/python-decompile3/files/5471266/pycfile.zip)
[pydisasm.zip](https://github.com/rocky/python-decompile3/files/5471267/pydisasm.zip)
|
1.0
|
Parse error at or near `DUP_TOP' (python 3.8) - When I decompiled the pyc, got the following error messages.
OS : win 10
Python 3.8.6
uncompyle6 : 3.7.4
decompyle3 : 3.3.2
pydisasm version 5.0.4
Python bytecode 3.8 (3413)
Source code size mod 2**32
```
Instruction context:
L. 5 52 LOAD_NAME list
54 LOAD_GENEXPR '<code_object <genexpr>>'
56 LOAD_STR '<genexpr>'
58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
60 LOAD_GENEXPR '<code_object <genexpr>>'
62 LOAD_STR '<genexpr>'
64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
66 LOAD_NAME f_in
68 GET_ITER
70 CALL_FUNCTION_1 1 ''
72 GET_ITER
74 CALL_FUNCTION_1 1 ''
76 CALL_FUNCTION_1 1 ''
78 STORE_NAME requirements
80 POP_BLOCK
82 LOAD_CONST None
-> 84 DUP_TOP
86 DUP_TOP
88 CALL_FUNCTION_3 3 ''
90 POP_TOP
92 JUMP_FORWARD 110 'to 110'
94_0 COME_FROM_WITH 48 '48'
94 <49>
96 POP_JUMP_IF_TRUE 100 'to 100'
98 <48>
100_0 COME_FROM 96 '96'
100 POP_TOP
102 POP_TOP
104 POP_TOP
106 POP_EXCEPT
108 POP_TOP
110_0 COME_FROM 92 '92'
# file start.pyc
# --- This code section failed: ---
L. 2 0 LOAD_CONST 0
2 LOAD_CONST None
4 IMPORT_NAME re
6 STORE_NAME re
8 LOAD_CONST 0
10 LOAD_CONST None
12 IMPORT_NAME os
14 STORE_NAME os
16 LOAD_CONST 0
18 LOAD_CONST None
20 IMPORT_NAME sys
22 STORE_NAME sys
24 LOAD_CONST 0
26 LOAD_CONST None
28 IMPORT_NAME pkg_resources
30 STORE_NAME pkg_resources
32 LOAD_CONST 0
34 LOAD_CONST None
36 IMPORT_NAME random
38 STORE_NAME random
L. 3 40 SETUP_FINALLY 228 'to 228'
L. 4 42 LOAD_NAME open
44 LOAD_STR 'requirements.txt'
46 CALL_FUNCTION_1 1 ''
48 SETUP_WITH 94 'to 94'
50 STORE_NAME f_in
L. 5 52 LOAD_NAME list
54 LOAD_GENEXPR '<code_object <genexpr>>'
56 LOAD_STR '<genexpr>'
58 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
60 LOAD_GENEXPR '<code_object <genexpr>>'
62 LOAD_STR '<genexpr>'
64 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
66 LOAD_NAME f_in
68 GET_ITER
70 CALL_FUNCTION_1 1 ''
72 GET_ITER
74 CALL_FUNCTION_1 1 ''
76 CALL_FUNCTION_1 1 ''
78 STORE_NAME requirements
80 POP_BLOCK
82 LOAD_CONST None
84 DUP_TOP
86 DUP_TOP
88 CALL_FUNCTION_3 3 ''
90 POP_TOP
92 JUMP_FORWARD 110 'to 110'
94_0 COME_FROM_WITH 48 '48'
94 <49>
96 POP_JUMP_IF_TRUE 100 'to 100'
98 <48>
100_0 COME_FROM 96 '96'
100 POP_TOP
102 POP_TOP
104 POP_TOP
106 POP_EXCEPT
108 POP_TOP
110_0 COME_FROM 92 '92'
L. 6 110 LOAD_NAME requirements
112 GET_ITER
114_0 COME_FROM 222 '222'
114_1 COME_FROM 210 '210'
114_2 COME_FROM 132 '132'
114 FOR_ITER 224 'to 224'
116 STORE_NAME require
L. 7 118 SETUP_FINALLY 134 'to 134'
L. 8 120 LOAD_NAME pkg_resources
122 LOAD_METHOD require
124 LOAD_NAME require
126 CALL_METHOD_1 1 ''
128 POP_TOP
130 POP_BLOCK
132 JUMP_BACK 114 'to 114'
134_0 COME_FROM_FINALLY 118 '118'
L. 9 134 DUP_TOP
136 LOAD_NAME Exception
138 <121> 220 ''
140 POP_TOP
142 STORE_NAME e
144 POP_TOP
146 SETUP_FINALLY 212 'to 212'
L. 10 148 LOAD_NAME re
150 LOAD_METHOD search
152 LOAD_STR "\\'(.*?)\\'"
154 LOAD_NAME str
156 LOAD_NAME e
158 CALL_FUNCTION_1 1 ''
160 CALL_METHOD_2 2 ''
162 LOAD_METHOD group
164 LOAD_CONST 1
166 CALL_METHOD_1 1 ''
168 STORE_NAME a
L. 11 170 LOAD_NAME print
172 LOAD_STR 'Installing '
174 LOAD_NAME a
176 BINARY_ADD
178 LOAD_STR ' ...'
180 BINARY_ADD
182 CALL_FUNCTION_1 1 ''
184 POP_TOP
L. 12 186 LOAD_NAME os
188 LOAD_METHOD system
190 LOAD_STR 'pip install '
192 LOAD_NAME a
194 BINARY_ADD
196 CALL_METHOD_1 1 ''
198 POP_TOP
200 POP_BLOCK
202 POP_EXCEPT
204 LOAD_CONST None
206 STORE_NAME e
208 DELETE_NAME e
210 JUMP_BACK 114 'to 114'
212_0 COME_FROM_FINALLY 146 '146'
212 LOAD_CONST None
214 STORE_NAME e
216 DELETE_NAME e
218 <48>
220 <48>
222 JUMP_BACK 114 'to 114'
224_0 COME_FROM 114 '114'
224 POP_BLOCK
226 JUMP_FORWARD 280 'to 280'
228_0 COME_FROM_FINALLY 40 '40'
L. 13 228 DUP_TOP
230 LOAD_NAME Exception
232_234 <121> 278 ''
236 POP_TOP
238 STORE_NAME e
240 POP_TOP
242 SETUP_FINALLY 270 'to 270'
L. 14 244 LOAD_NAME print
246 LOAD_STR 'installation error: '
248 LOAD_NAME e
250 FORMAT_VALUE 0 ''
252 BUILD_STRING_2 2
254 CALL_FUNCTION_1 1 ''
256 POP_TOP
258 POP_BLOCK
260 POP_EXCEPT
262 LOAD_CONST None
264 STORE_NAME e
266 DELETE_NAME e
268 JUMP_FORWARD 280 'to 280'
270_0 COME_FROM_FINALLY 242 '242'
270 LOAD_CONST None
272 STORE_NAME e
274 DELETE_NAME e
276 <48>
278 <48>
280_0 COME_FROM 268 '268'
280_1 COME_FROM 226 '226'
L. 16 280 LOAD_CONST 0
282 LOAD_CONST None
284 IMPORT_NAME asyncio
286 STORE_NAME asyncio
288 LOAD_CONST 0
290 LOAD_CONST None
292 IMPORT_NAME logging
294 STORE_NAME logging
296 LOAD_CONST 0
298 LOAD_CONST None
300 IMPORT_NAME requests
302 STORE_NAME requests
304 LOAD_CONST 0
306 LOAD_CONST None
308 IMPORT_NAME sqlite3
310 STORE_NAME sqlite3
312 LOAD_CONST 0
314 LOAD_CONST None
316 IMPORT_NAME base64
318 STORE_NAME base64
L. 17 320 LOAD_CONST 0
322 LOAD_CONST ('sleep', 'strptime')
324 IMPORT_NAME time
326 IMPORT_FROM sleep
328 STORE_NAME sleep
330 IMPORT_FROM strptime
332 STORE_NAME strptime
334 POP_TOP
L. 18 336 LOAD_CONST 0
338 LOAD_CONST ('KeyboardButtonUrl', 'MessageMediaContact', 'ReplyInlineMarkup', 'UpdateShortMessage')
340 IMPORT_NAME_ATTR telethon.tl.types
342 IMPORT_FROM KeyboardButtonUrl
344 STORE_NAME KeyboardButtonUrl
346 IMPORT_FROM MessageMediaContact
348 STORE_NAME MessageMediaContact
350 IMPORT_FROM ReplyInlineMarkup
352 STORE_NAME ReplyInlineMarkup
354 IMPORT_FROM UpdateShortMessage
356 STORE_NAME UpdateShortMessage
358 POP_TOP
L. 19 360 LOAD_CONST 0
362 LOAD_CONST ('TelegramClient', 'client', 'errors', 'events', 'functions', 'connection')
364 IMPORT_NAME telethon
366 IMPORT_FROM TelegramClient
368 STORE_NAME TelegramClient
370 IMPORT_FROM client
372 STORE_GLOBAL client
374 IMPORT_FROM errors
376 STORE_NAME errors
378 IMPORT_FROM events
380 STORE_NAME events
382 IMPORT_FROM functions
384 STORE_NAME functions
386 IMPORT_FROM connection
388 STORE_NAME connection
390 POP_TOP
L. 20 392 LOAD_CONST 0
394 LOAD_CONST ('JoinChannelRequest',)
396 IMPORT_NAME_ATTR telethon.tl.functions.channels
398 IMPORT_FROM JoinChannelRequest
400 STORE_NAME JoinChannelRequest
402 POP_TOP
L. 21 404 LOAD_CONST 0
406 LOAD_CONST ('GetBotCallbackAnswerRequest', 'GetHistoryRequest', 'ImportChatInviteRequest', 'GetMessagesViewsRequest')
408 IMPORT_NAME_ATTR telethon.tl.functions.messages
410 IMPORT_FROM GetBotCallbackAnswerRequest
412 STORE_NAME GetBotCallbackAnswerRequest
414 IMPORT_FROM GetHistoryRequest
416 STORE_NAME GetHistoryRequest
418 IMPORT_FROM ImportChatInviteRequest
420 STORE_NAME ImportChatInviteRequest
422 IMPORT_FROM GetMessagesViewsRequest
424 STORE_NAME GetMessagesViewsRequest
426 POP_TOP
L. 22 428 LOAD_CONST 0
430 LOAD_CONST ('datetime', 'timedelta')
432 IMPORT_NAME datetime
434 IMPORT_FROM datetime
436 STORE_NAME datetime
438 IMPORT_FROM timedelta
440 STORE_NAME timedelta
442 POP_TOP
L. 23 444 LOAD_CONST 0
446 LOAD_CONST ('Fore', 'Style', 'init')
448 IMPORT_NAME colorama
450 IMPORT_FROM Fore
452 STORE_NAME Fore
454 IMPORT_FROM Style
456 STORE_NAME Style
458 IMPORT_FROM init
460 STORE_NAME color_ama
462 POP_TOP
L. 24 464 LOAD_CONST 0
466 LOAD_CONST ('BeautifulSoup',)
468 IMPORT_NAME bs4
470 IMPORT_FROM BeautifulSoup
472 STORE_NAME BeautifulSoup
474 POP_TOP
L. 25 476 LOAD_CONST 0
478 LOAD_CONST None
480 IMPORT_NAME random
482 STORE_NAME random
L. 26 484 LOAD_CONST 0
486 LOAD_CONST None
488 IMPORT_NAME socks
490 STORE_NAME socks
L. 27 492 LOAD_NAME logging
494 LOAD_ATTR basicConfig
496 LOAD_NAME logging
498 LOAD_ATTR ERROR
500 LOAD_CONST ('level',)
502 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args'
504 POP_TOP
L. 28 506 LOAD_NAME color_ama
508 LOAD_CONST True
510 LOAD_CONST ('autoreset',)
512 CALL_FUNCTION_KW_1 1 '1 total positional and keyword args'
514 POP_TOP
L. 29 516 LOAD_NAME os
518 LOAD_METHOD system
520 LOAD_NAME os
522 LOAD_ATTR name
524 LOAD_STR 'nt'
526 COMPARE_OP ==
528_530 POP_JUMP_IF_FALSE 536 'to 536'
532 LOAD_STR 'cls'
534 JUMP_FORWARD 538 'to 538'
536_0 COME_FROM 528 '528'
536 LOAD_STR 'clear'
538_0 COME_FROM 534 '534'
538 CALL_METHOD_1 1 ''
540 POP_TOP
L. 30 542 LOAD_STR '\n'
L. 31 544 LOAD_NAME Style
546 LOAD_ATTR NORMAL
548 LOAD_NAME Fore
550 LOAD_ATTR MAGENTA
552 BINARY_ADD
L. 30 554 FORMAT_VALUE 0 ''
556 LOAD_STR ' ____ ___ ___ _ _ '
L. 31 558 LOAD_NAME Fore
560 LOAD_ATTR GREEN
L. 30 562 FORMAT_VALUE 0 ''
564 LOAD_STR ' ___ ___ \n'
L. 31 566 LOAD_NAME Style
568 LOAD_ATTR NORMAL
570 LOAD_NAME Fore
572 LOAD_ATTR MAGENTA
574 BINARY_ADD
L. 30 576 FORMAT_VALUE 0 ''
578 LOAD_STR ' / ___| / _ \\ |_ _| | \\ | | '
L. 31 580 LOAD_NAME Fore
582 LOAD_ATTR GREEN
L. 30 584 FORMAT_VALUE 0 ''
586 LOAD_STR ' / _ \\ ( _ ) \n'
L. 31 588 LOAD_NAME Style
590 LOAD_ATTR NORMAL
592 LOAD_NAME Fore
594 LOAD_ATTR MAGENTA
596 BINARY_ADD
L. 30 598 FORMAT_VALUE 0 ''
600 LOAD_STR ' | | | | | | | | | \\| | '
L. 31 602 LOAD_NAME Fore
604 LOAD_ATTR GREEN
L. 30 606 FORMAT_VALUE 0 ''
608 LOAD_STR '| (_) | / _ \\ \n'
L. 31 610 LOAD_NAME Style
612 LOAD_ATTR NORMAL
614 LOAD_NAME Fore
616 LOAD_ATTR MAGENTA
618 BINARY_ADD
L. 30 620 FORMAT_VALUE 0 ''
622 LOAD_STR ' | |___ | |_| | | | | |\\ | '
L. 31 624 LOAD_NAME Fore
626 LOAD_ATTR GREEN
L. 30 628 FORMAT_VALUE 0 ''
630 LOAD_STR ' \\__, | | (_) |\n'
L. 31 632 LOAD_NAME Style
634 LOAD_ATTR NORMAL
636 LOAD_NAME Fore
638 LOAD_ATTR MAGENTA
640 BINARY_ADD
L. 30 642 FORMAT_VALUE 0 ''
644 LOAD_STR ' \\____| \\___/ |___| |_| \\_| '
L. 31 646 LOAD_NAME Fore
648 LOAD_ATTR GREEN
L. 30 650 FORMAT_VALUE 0 ''
652 LOAD_STR ' /_/ \\___/ \n'
L. 36 654 LOAD_NAME Fore
656 LOAD_ATTR BLUE
L. 30 658 FORMAT_VALUE 0 ''
660 LOAD_STR ' Edit By'
L. 36 662 LOAD_NAME Style
664 LOAD_ATTR DIM
666 LOAD_NAME Fore
668 LOAD_ATTR RED
670 BINARY_ADD
L. 30 672 FORMAT_VALUE 0 ''
674 LOAD_STR ':'
L. 36 676 LOAD_NAME Fore
678 LOAD_ATTR YELLOW
L. 30 680 FORMAT_VALUE 0 ''
682 LOAD_STR ' Abbas Bachari '
L. 36 684 LOAD_NAME Fore
686 LOAD_ATTR BLUE
L. 30 688 FORMAT_VALUE 0 ''
690 LOAD_STR 'Version'
L. 36 692 LOAD_NAME Fore
694 LOAD_ATTR YELLOW
L. 30 696 FORMAT_VALUE 0 ''
698 LOAD_STR ' 7.9\n'
L. 37 700 LOAD_NAME Style
702 LOAD_ATTR NORMAL
704 LOAD_NAME Fore
706 LOAD_ATTR RED
708 BINARY_ADD
L. 30 710 FORMAT_VALUE 0 ''
712 LOAD_STR '==============================================\n'
L. 38 714 LOAD_NAME Style
716 LOAD_ATTR BRIGHT
718 LOAD_NAME Fore
720 LOAD_ATTR GREEN
722 BINARY_ADD
L. 30 724 FORMAT_VALUE 0 ''
726 LOAD_STR 'Sponsor Channel '
L. 36 728 LOAD_NAME Style
730 LOAD_ATTR DIM
732 LOAD_NAME Fore
734 LOAD_ATTR RED
736 BINARY_ADD
L. 30 738 FORMAT_VALUE 0 ''
740 LOAD_STR ':'
L. 38 742 LOAD_NAME Style
744 LOAD_ATTR RESET_ALL
L. 30 746 FORMAT_VALUE 0 ''
748 LOAD_STR ' @COIN98'
750 BUILD_STRING_39 39
752 STORE_NAME banner
L. 41 754 LOAD_CONST 799906641
756 LOAD_STR 'BitcoinClick_bot'
758 LOAD_STR '/start 2fv0'
760 LOAD_CONST ('id', 'username', 'start')
762 BUILD_CONST_KEY_MAP_3 3
L. 42 764 LOAD_CONST 741849360
766 LOAD_STR 'Litecoin_click_bot'
768 LOAD_STR '/start aZYG'
770 LOAD_CONST ('id', 'username', 'start')
772 BUILD_CONST_KEY_MAP_3 3
L. 43 774 LOAD_CONST 715510199
776 LOAD_STR 'Dogecoin_click_bot'
778 LOAD_STR '/start ljSP'
780 LOAD_CONST ('id', 'username', 'start')
782 BUILD_CONST_KEY_MAP_3 3
L. 44 784 LOAD_CONST 687127269
786 LOAD_STR 'BCH_clickbot'
788 LOAD_STR '/start 3su3'
790 LOAD_CONST ('id', 'username', 'start')
792 BUILD_CONST_KEY_MAP_3 3
L. 45 794 LOAD_CONST 850081470
796 LOAD_STR 'Zcash_click_bot'
798 LOAD_STR '/start 6qKG'
800 LOAD_CONST ('id', 'username', 'start')
802 BUILD_CONST_KEY_MAP_3 3
L. 40 804 LOAD_CONST ('BTC', 'LTC', 'DOGE', 'BCH', 'ZEC')
806 BUILD_CONST_KEY_MAP_5 5
808 STORE_NAME Symbols
L. 48 810 LOAD_CONST 197597
812 STORE_GLOBAL api_id
L. 49 814 LOAD_STR '227647fae1d2a3a1419bdd527337c87d'
816 STORE_GLOBAL api_hash
L. 50 818 LOAD_STR 'RGVhciB1c2VyLCB5b3UgbmVlZCB0byBzdWJzY3JpYmUgdG8gdGhlCiAgICBmb2xsb3dpbmcgVGVsZWdyYW0gY2hhbm5lbBtbMTswbQogICAgQ2hhbm5lbCBJRCA6IEBDT0lOOTgNCgoK'
820 STORE_NAME MSG
L. 51 822 LOAD_NAME requests
824 LOAD_METHOD session
826 CALL_METHOD_0 0 ''
828 STORE_GLOBAL session
L. 52 830 LOAD_STR 'User-Agent'
832 LOAD_STR 'Mozilla/5.0 (Linux; Android 9; moto g(7) play) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.111 Mobile Safari/537.36'
834 BUILD_MAP_1 1
836 STORE_NAME ua
L. 53 838 LOAD_STR ''
840 STORE_GLOBAL dogeclick_channel
L. 54 842 LOAD_STR ''
844 STORE_GLOBAL phone_number
L. 55 846 LOAD_STR ''
848 STORE_GLOBAL Wallet
L. 56 850 LOAD_STR ''
852 STORE_GLOBAL isVisit
L. 57 854 LOAD_STR ''
856 STORE_GLOBAL isJoin
L. 58 858 LOAD_STR ''
860 STORE_GLOBAL isBotMs
L. 59 862 LOAD_STR ''
864 STORE_GLOBAL isLoop
L. 60 866 LOAD_STR ''
868 STORE_GLOBAL Balance
L. 61 870 BUILD_MAP_0 0
872 STORE_GLOBAL SET
L. 62 874 LOAD_STR ''
876 STORE_GLOBAL coin
L. 63 878 LOAD_CONST 0
880 STORE_NAME INDEX
L. 64 882 LOAD_CONST ()
884 STORE_GLOBAL NUM
L. 65 886 LOAD_STR 'SIT'
888 STORE_GLOBAL stat
L. 66 890 LOAD_CONST None
892 STORE_GLOBAL captcha
L. 67 894 LOAD_CONST None
896 STORE_GLOBAL sit
L. 68 898 LOAD_STR ''
900 STORE_GLOBAL message
L. 69 902 LOAD_STR ''
904 STORE_GLOBAL channel_name
L. 70 906 LOAD_CONST None
908 STORE_GLOBAL channel_id
L. 71 910 LOAD_STR ''
912 STORE_GLOBAL bot_name
L. 72 914 LOAD_STR ''
916 STORE_GLOBAL COMENT
L. 73 918 LOAD_CONST 7.85
920 STORE_NAME APVER
L. 74 922 BUILD_LIST_0 0
924 STORE_GLOBAL BOTS
L. 75 926 BUILD_LIST_0 0
928 STORE_GLOBAL CHNS
L. 76 930 BUILD_LIST_0 0
932 STORE_GLOBAL GRPS
L. 77 934 BUILD_LIST_0 0
936 STORE_GLOBAL USRS
L. 78 938 LOAD_CONST 0
940 STORE_GLOBAL LIMIT
L. 79 942 LOAD_STR ''
944 STORE_GLOBAL LICE
L. 80 946 LOAD_STR ''
948 STORE_GLOBAL MASE
L. 81 950 LOAD_CONST 0
952 STORE_GLOBAL VER
L. 82 954 LOAD_CONST 0
956 STORE_GLOBAL mass_id
L. 83 958 LOAD_CONST None
960 STORE_GLOBAL click_data
L. 84 962 LOAD_STR ''
964 STORE_GLOBAL VIPMES
L. 85 966 LOAD_CONST 0
968 STORE_GLOBAL JNUM
L. 86 970 LOAD_CONST 0
972 STORE_GLOBAL ERN
L. 87 974 LOAD_CONST 0
976 STORE_GLOBAL WIWNUM
L. 88 978 LOAD_CONST 0
980 STORE_GLOBAL WIWCON
L. 89 982 LOAD_CONST 0
984 STORE_NAME LOPS
L. 90 986 BUILD_MAP_0 0
988 STORE_GLOBAL SETIN
L. 91 990 BUILD_LIST_0 0
992 STORE_GLOBAL PLOGINS
L. 92 994 BUILD_MAP_0 0
996 STORE_GLOBAL MESEGS
L. 93 998 BUILD_MAP_0 0
1000 STORE_GLOBAL ADMESEG
L. 94 1002 LOAD_STR 'YES'
1004 STORE_GLOBAL ADD
L. 95 1006 LOAD_CONST None
1008 STORE_GLOBAL MINWD
L. 96 1010 LOAD_STR 'NUMBERS.txt'
1012 STORE_GLOBAL path
L. 97 1014 LOAD_CONST None
1016 STORE_GLOBAL Proxy
L. 99 1018 LOAD_CODE <code_object isNumber>
1020 LOAD_STR 'isNumber'
1022 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1024 STORE_NAME isNumber
L. 105 1026 LOAD_CODE <code_object request>
1028 LOAD_STR 'request'
1030 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1032 STORE_NAME request
L. 131 1034 LOAD_CODE <code_object GetIP>
1036 LOAD_STR 'GetIP'
1038 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1040 STORE_NAME GetIP
L. 145 1042 LOAD_CODE <code_object SEND_CONTACT>
1044 LOAD_STR 'SEND_CONTACT'
1046 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1048 STORE_NAME SEND_CONTACT
L. 159 1050 LOAD_CODE <code_object send_messege_to_members>
1052 LOAD_STR 'send_messege_to_members'
1054 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1056 STORE_NAME send_messege_to_members
L. 203 1058 LOAD_CODE <code_object CHNUM>
1060 LOAD_STR 'CHNUM'
1062 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1064 STORE_NAME CHNUM
L. 211 1066 LOAD_CODE <code_object numbers>
1068 LOAD_STR 'numbers'
1070 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1072 STORE_NAME numbers
L. 272 1074 LOAD_CODE <code_object AnonsRobot>
1076 LOAD_STR 'AnonsRobot'
1078 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1080 STORE_NAME AnonsRobot
L. 285 1082 LOAD_CODE <code_object START_PLOGINS>
1084 LOAD_STR 'START_PLOGINS'
1086 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1088 STORE_NAME START_PLOGINS
L. 382 1090 LOAD_CODE <code_object SEND_PLOGIN_MESSEGE>
1092 LOAD_STR 'SEND_PLOGIN_MESSEGE'
1094 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1096 STORE_NAME SEND_PLOGIN_MESSEGE
L. 403 1098 LOAD_CODE <code_object SEND_GROUP_MESSEGE>
1100 LOAD_STR 'SEND_GROUP_MESSEGE'
1102 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1104 STORE_NAME SEND_GROUP_MESSEGE
L. 435 1106 LOAD_CODE <code_object Settings>
1108 LOAD_STR 'Settings'
1110 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1112 STORE_NAME Settings
L. 569 1114 LOAD_CODE <code_object SEND_TO_PRIVAIT_GROUP>
1116 LOAD_STR 'SEND_TO_PRIVAIT_GROUP'
1118 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1120 STORE_NAME SEND_TO_PRIVAIT_GROUP
L. 611 1122 LOAD_CODE <code_object Creat_user>
1124 LOAD_STR 'Creat_user'
1126 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1128 STORE_NAME Creat_user
L. 617 1130 LOAD_CODE <code_object Add_Plogin>
1132 LOAD_STR 'Add_Plogin'
1134 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1136 STORE_NAME Add_Plogin
L. 629 1138 LOAD_CODE <code_object Delete_Plogin>
1140 LOAD_STR 'Delete_Plogin'
1142 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1144 STORE_NAME Delete_Plogin
L. 635 1146 LOAD_CODE <code_object GetLoop>
1148 LOAD_STR 'GetLoop'
1150 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1152 STORE_NAME GetLoop
L. 648 1154 LOAD_CODE <code_object Add_Channel>
1156 LOAD_STR 'Add_Channel'
1158 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1160 STORE_NAME Add_Channel
L. 663 1162 LOAD_STR 'NO'
1164 STORE_GLOBAL ISWID
L. 664 1166 LOAD_CODE <code_object Delete_Cannels>
1168 LOAD_STR 'Delete_Cannels'
1170 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1172 STORE_NAME Delete_Cannels
L. 711 1174 LOAD_CODE <code_object reset>
1176 LOAD_STR 'reset'
1178 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1180 STORE_NAME reset
L. 736 1182 LOAD_CODE <code_object maseg>
1184 LOAD_STR 'maseg'
1186 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1188 STORE_NAME maseg
L. 740 1190 LOAD_CODE <code_object WAIT>
1192 LOAD_STR 'WAIT'
1194 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1196 STORE_NAME WAIT
L. 777 1198 LOAD_CODE <code_object FloodTimer>
1200 LOAD_STR 'FloodTimer'
1202 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1204 STORE_NAME FloodTimer
L. 787 1206 LOAD_CODE <code_object sleepTime>
1208 LOAD_STR 'sleepTime'
1210 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1212 STORE_NAME sleepTime
L. 806 1214 LOAD_CODE <code_object JoinTimer>
1216 LOAD_STR 'JoinTimer'
1218 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1220 STORE_NAME JoinTimer
L. 816 1222 LOAD_CODE <code_object GET_USERNAME>
1224 LOAD_STR 'GET_USERNAME'
1226 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1228 STORE_NAME GET_USERNAME
L. 839 1230 LOAD_CODE <code_object forward_messages>
1232 LOAD_STR 'forward_messages'
1234 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1236 STORE_NAME forward_messages
L. 854 1238 LOAD_CODE <code_object start>
1240 LOAD_STR 'start'
1242 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1244 STORE_NAME start
L. 997 1246 LOAD_CODE <code_object main>
1248 LOAD_STR 'main'
1250 MAKE_FUNCTION_0 'Neither defaults, keyword-only args, annotations, nor closures'
1252 STORE_NAME main
1254_0 COME_FROM 1756 '1756'
1254_1 COME_FROM 1752 '1752'
1254_2 COME_FROM 1718 '1718'
L.1293 1254_1256 SETUP_FINALLY 1720 'to 1720'
L.1294 1258 LOAD_NAME asyncio
1260 LOAD_METHOD get_event_loop
1262 CALL_METHOD_0 0 ''
1264 LOAD_METHOD run_until_complete
1266 LOAD_NAME numbers
1268 CALL_FUNCTION_0 0 ''
1270 CALL_METHOD_1 1 ''
1272 POP_TOP
L.1295 1274 LOAD_NAME LOPS
1276 LOAD_CONST 1
1278 INPLACE_ADD
1280 STORE_NAME LOPS
L.1296 1282 LOAD_GLOBAL NUM
1284 GET_ITER
1286_0 COME_FROM 1328 '1328'
1286 FOR_ITER 1332 'to 1332'
1288 STORE_NAME i
L.1297 1290 LOAD_NAME INDEX
1292 LOAD_CONST 1
1294 INPLACE_ADD
1296 STORE_NAME INDEX
L.1298 1298 LOAD_CONST 0
1300 STORE_GLOBAL JNUM
L.1299 1302 LOAD_NAME asyncio
1304 LOAD_METHOD get_event_loop
1306 CALL_METHOD_0 0 ''
1308 LOAD_METHOD run_until_complete
1310 LOAD_NAME main
1312 LOAD_NAME i
1314 CALL_FUNCTION_1 1 ''
1316 CALL_METHOD_1 1 ''
1318 POP_TOP
L.1302 1320 LOAD_NAME sleep
1322 LOAD_CONST 3
1324 CALL_FUNCTION_1 1 ''
1326 POP_TOP
1328_1330 JUMP_BACK 1286 'to 1286'
1332_0 COME_FROM 1286 '1286'
L.1303 1332 LOAD_NAME os
1334 LOAD_METHOD system
1336 LOAD_NAME os
1338 LOAD_ATTR name
1340 LOAD_STR 'nt'
1342 COMPARE_OP ==
1344_1346 POP_JUMP_IF_FALSE 1352 'to 1352'
1348 LOAD_STR 'cls'
1350 JUMP_FORWARD 1354 'to 1354'
1352_0 COME_FROM 1344 '1344'
1352 LOAD_STR 'clear'
1354_0 COME_FROM 1350 '1350'
1354 CALL_METHOD_1 1 ''
1356 POP_TOP
L.1304 1358 LOAD_NAME format
1360 LOAD_GLOBAL ERN
1362 LOAD_STR '.8f'
1364 CALL_FUNCTION_2 2 ''
1366 STORE_NAME ER
L.1305 1368 LOAD_NAME format
1370 LOAD_GLOBAL WIWCON
1372 LOAD_STR '.8f'
1374 CALL_FUNCTION_2 2 ''
1376 STORE_NAME WC
L.1306 1378 LOAD_GLOBAL LICE
1380 LOAD_STR 'VIP'
1382 COMPARE_OP ==
1384_1386 POP_JUMP_IF_FALSE 1406 'to 1406'
L.1307 1388 LOAD_GLOBAL LICE
1390 FORMAT_VALUE 0 ''
1392 LOAD_STR '('
1394 LOAD_GLOBAL LIMIT
1396 FORMAT_VALUE 0 ''
1398 LOAD_STR ')'
1400 BUILD_STRING_4 4
1402 STORE_NAME VI
1404 JUMP_FORWARD 1428 'to 1428'
1406_0 COME_FROM 1384 '1384'
L.1309 1406 LOAD_NAME Fore
1408 LOAD_ATTR RED
1410 FORMAT_VALUE 0 ''
1412 LOAD_GLOBAL LICE
1414 FORMAT_VALUE 0 ''
1416 LOAD_STR '('
1418 LOAD_GLOBAL LIMIT
1420 FORMAT_VALUE 0 ''
1422 LOAD_STR ')'
1424 BUILD_STRING_5 5
1426 STORE_NAME VI
1428_0 COME_FROM 1404 '1404'
L.1311 1428 LOAD_NAME print
1430 LOAD_NAME banner
1432 CALL_FUNCTION_1 1 ''
1434 POP_TOP
L.1312 1436 LOAD_NAME print
1438 LOAD_NAME Fore
1440 LOAD_ATTR GREEN
1442 FORMAT_VALUE 0 ''
1444 LOAD_STR 'Script License '
1446 LOAD_NAME Fore
1448 LOAD_ATTR RED
1450 FORMAT_VALUE 0 ''
1452 LOAD_STR ':'
1454 LOAD_NAME Fore
1456 LOAD_ATTR RESET
1458 FORMAT_VALUE 0 ''
1460 LOAD_STR ' '
1462 LOAD_NAME VI
1464 FORMAT_VALUE 0 ''
1466 BUILD_STRING_7 7
1468 CALL_FUNCTION_1 1 ''
1470 POP_TOP
L.1313 1472 LOAD_NAME print
1474 LOAD_NAME Fore
1476 LOAD_ATTR GREEN
1478 FORMAT_VALUE 0 ''
1480 LOAD_STR 'Total Loops '
1482 LOAD_NAME Fore
1484 LOAD_ATTR RED
1486 FORMAT_VALUE 0 ''
1488 LOAD_STR ':'
1490 LOAD_NAME Fore
1492 LOAD_ATTR RESET
1494 FORMAT_VALUE 0 ''
1496 LOAD_STR ' '
1498 LOAD_NAME LOPS
1500 FORMAT_VALUE 0 ''
1502 BUILD_STRING_7 7
1504 CALL_FUNCTION_1 1 ''
1506 POP_TOP
L.1314 1508 LOAD_NAME print
1510 LOAD_NAME Fore
1512 LOAD_ATTR GREEN
1514 FORMAT_VALUE 0 ''
1516 LOAD_STR 'Total accounts '
1518 LOAD_NAME Fore
1520 LOAD_ATTR RED
1522 FORMAT_VALUE 0 ''
1524 LOAD_STR ':'
1526 LOAD_NAME Fore
1528 LOAD_ATTR RESET
1530 FORMAT_VALUE 0 ''
1532 LOAD_STR ' '
1534 LOAD_NAME len
1536 LOAD_GLOBAL NUM
1538 CALL_FUNCTION_1 1 ''
1540 FORMAT_VALUE 0 ''
1542 BUILD_STRING_7 7
1544 CALL_FUNCTION_1 1 ''
1546 POP_TOP
L.1315 1548 LOAD_NAME print
1550 LOAD_NAME Fore
1552 LOAD_ATTR GREEN
1554 FORMAT_VALUE 0 ''
1556 LOAD_STR 'accounts withdraw '
1558 LOAD_NAME Fore
1560 LOAD_ATTR RED
1562 FORMAT_VALUE 0 ''
1564 LOAD_STR ':'
1566 LOAD_NAME Fore
1568 LOAD_ATTR RESET
1570 FORMAT_VALUE 0 ''
1572 LOAD_STR ' '
1574 LOAD_GLOBAL WIWNUM
1576 FORMAT_VALUE 0 ''
1578 LOAD_STR ' account'
1580 BUILD_STRING_8 8
1582 CALL_FUNCTION_1 1 ''
1584 POP_TOP
L.1316 1586 LOAD_NAME print
1588 LOAD_NAME Fore
1590 LOAD_ATTR GREEN
1592 FORMAT_VALUE 0 ''
1594 LOAD_STR 'Total withdraw '
1596 LOAD_NAME Fore
1598 LOAD_ATTR RED
1600 FORMAT_VALUE 0 ''
1602 LOAD_STR ':'
1604 LOAD_NAME Fore
1606 LOAD_ATTR RESET
1608 FORMAT_VALUE 0 ''
1610 LOAD_STR ' '
1612 LOAD_NAME WC
1614 FORMAT_VALUE 0 ''
1616 LOAD_STR ' '
1618 LOAD_GLOBAL coin
1620 FORMAT_VALUE 0 ''
1622 BUILD_STRING_9 9
1624 CALL_FUNCTION_1 1 ''
1626 POP_TOP
L.1317 1628 LOAD_NAME print
1630 LOAD_NAME Fore
1632 LOAD_ATTR GREEN
1634 FORMAT_VALUE 0 ''
1636 LOAD_STR 'Total balance '
1638 LOAD_NAME Fore
1640 LOAD_ATTR RED
1642 FORMAT_VALUE 0 ''
1644 LOAD_STR ':'
1646 LOAD_NAME Fore
1648 LOAD_ATTR RESET
1650 FORMAT_VALUE 0 ''
1652 LOAD_STR ' '
1654 LOAD_NAME ER
1656 FORMAT_VALUE 0 ''
1658 LOAD_STR ' '
1660 LOAD_GLOBAL coin
1662 FORMAT_VALUE 0 ''
1664 BUILD_STRING_9 9
1666 CALL_FUNCTION_1 1 ''
1668 POP_TOP
L.1318 1670 LOAD_NAME print
1672 LOAD_STR '\n'
1674 CALL_FUNCTION_1 1 ''
1676 POP_TOP
L.1319 1678 LOAD_GLOBAL SET
1680 LOAD_STR 'AUTOLOOP'
1682 BINARY_SUBSCR
1684 LOAD_METHOD upper
1686 CALL_METHOD_0 0 ''
1688 LOAD_STR 'YES'
1690 COMPARE_OP !=
1692_1694 POP_JUMP_IF_FALSE 1702 'to 1702'
L.1320 1696 POP_BLOCK
1698_1700 BREAK_LOOP 1760 'to 1760'
1702_0 COME_FROM 1692 '1692'
L.1322 1702 LOAD_NAME sleepTime
1704 CALL_FUNCTION_0 0 ''
1706 POP_TOP
L.1323 1708 LOAD_CONST 0
1710 STORE_GLOBAL ERN
L.1324 1712 LOAD_CONST 0
1714 STORE_NAME INDEX
1716 POP_BLOCK
1718 JUMP_BACK 1254 'to 1254'
1720_0 COME_FROM_FINALLY 1254 '1254'
L.1326 1720 DUP_TOP
1722 LOAD_NAME KeyboardInterrupt
1724_1726 <121> 1754 ''
1728 POP_TOP
1730 POP_TOP
1732 POP_TOP
L.1327 1734 LOAD_NAME print
1736 LOAD_STR ''
1738 CALL_FUNCTION_1 1 ''
1740 POP_TOP
L.1328 1742 LOAD_NAME sys
1744 LOAD_METHOD exit
1746 CALL_METHOD_0 0 ''
1748 POP_TOP
1750 POP_EXCEPT
1752 JUMP_BACK 1254 'to 1254'
1754 <48>
1756_1758 JUMP_BACK 1254 'to 1254'
1760_0 COME_FROM 1698 '1698'
Parse error at or near `DUP_TOP' instruction at offset 84
```
i also attached : .`pyc `file , decompiled results and `pydisasm `results
[decompiled_output.zip](https://github.com/rocky/python-decompile3/files/5471265/decompiled_output.zip)
[pycfile.zip](https://github.com/rocky/python-decompile3/files/5471266/pycfile.zip)
[pydisasm.zip](https://github.com/rocky/python-decompile3/files/5471267/pydisasm.zip)
|
non_process
|
parse error at or near dup top python when i decompiled the pyc got the following error messages os win python pydisasm version python bytecode source code size mod instruction context l load name list load genexpr load str make function neither defaults keyword only args annotations nor closures load genexpr load str make function neither defaults keyword only args annotations nor closures load name f in get iter call function get iter call function call function store name requirements pop block load const none dup top dup top call function pop top jump forward to come from with pop jump if true to come from pop top pop top pop top pop except pop top come from file start pyc this code section failed l load const load const none import name re store name re load const load const none import name os store name os load const load const none import name sys store name sys load const load const none import name pkg resources store name pkg resources load const load const none import name random store name random l setup finally to l load name open load str requirements txt call function setup with to store name f in l load name list load genexpr load str make function neither defaults keyword only args annotations nor closures load genexpr load str make function neither defaults keyword only args annotations nor closures load name f in get iter call function get iter call function call function store name requirements pop block load const none dup top dup top call function pop top jump forward to come from with pop jump if true to come from pop top pop top pop top pop except pop top come from l load name requirements get iter come from come from come from for iter to store name require l setup finally to l load name pkg resources load method require load name require call method pop top pop block jump back to come from finally l dup top load name exception pop top store name e pop top setup finally to l load name re load method search load str load name str load name e call function call method load method group load const call method store name a l load name print load str installing load name a binary add load str binary add call function pop top l load name os load method system load str pip install load name a binary add call method pop top pop block pop except load const none store name e delete name e jump back to come from finally load const none store name e delete name e jump back to come from pop block jump forward to come from finally l dup top load name exception pop top store name e pop top setup finally to l load name print load str installation error load name e format value build string call function pop top pop block pop except load const none store name e delete name e jump forward to come from finally load const none store name e delete name e come from come from l load const load const none import name asyncio store name asyncio load const load const none import name logging store name logging load const load const none import name requests store name requests load const load const none import name store name load const load const none import name store name l load const load const sleep strptime import name time import from sleep store name sleep import from strptime store name strptime pop top l load const load const keyboardbuttonurl messagemediacontact replyinlinemarkup updateshortmessage import name attr telethon tl types import from keyboardbuttonurl store name keyboardbuttonurl import from messagemediacontact store name messagemediacontact import from replyinlinemarkup store name replyinlinemarkup import from updateshortmessage store name updateshortmessage pop top l load const load const telegramclient client errors events functions connection import name telethon import from telegramclient store name telegramclient import from client store global client import from errors store name errors import from events store name events import from functions store name functions import from connection store name connection pop top l load const load const joinchannelrequest import name attr telethon tl functions channels import from joinchannelrequest store name joinchannelrequest pop top l load const load const getbotcallbackanswerrequest gethistoryrequest importchatinviterequest getmessagesviewsrequest import name attr telethon tl functions messages import from getbotcallbackanswerrequest store name getbotcallbackanswerrequest import from gethistoryrequest store name gethistoryrequest import from importchatinviterequest store name importchatinviterequest import from getmessagesviewsrequest store name getmessagesviewsrequest pop top l load const load const datetime timedelta import name datetime import from datetime store name datetime import from timedelta store name timedelta pop top l load const load const fore style init import name colorama import from fore store name fore import from style store name style import from init store name color ama pop top l load const load const beautifulsoup import name import from beautifulsoup store name beautifulsoup pop top l load const load const none import name random store name random l load const load const none import name socks store name socks l load name logging load attr basicconfig load name logging load attr error load const level call function kw total positional and keyword args pop top l load name color ama load const true load const autoreset call function kw total positional and keyword args pop top l load name os load method system load name os load attr name load str nt compare op pop jump if false to load str cls jump forward to come from load str clear come from call method pop top l load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name style load attr normal load name fore load attr magenta binary add l format value load str l load name fore load attr green l format value load str n l load name fore load attr blue l format value load str edit by l load name style load attr dim load name fore load attr red binary add l format value load str l load name fore load attr yellow l format value load str abbas bachari l load name fore load attr blue l format value load str version l load name fore load attr yellow l format value load str n l load name style load attr normal load name fore load attr red binary add l format value load str n l load name style load attr bright load name fore load attr green binary add l format value load str sponsor channel l load name style load attr dim load name fore load attr red binary add l format value load str l load name style load attr reset all l format value load str build string store name banner l load const load str bitcoinclick bot load str start load const id username start build const key map l load const load str litecoin click bot load str start azyg load const id username start build const key map l load const load str dogecoin click bot load str start ljsp load const id username start build const key map l load const load str bch clickbot load str start load const id username start build const key map l load const load str zcash click bot load str start load const id username start build const key map l load const btc ltc doge bch zec build const key map store name symbols l load const store global api id l load str store global api hash l load str store name msg l load name requests load method session call method store global session l load str user agent load str mozilla linux android moto g play applewebkit khtml like gecko chrome mobile safari build map store name ua l load str store global dogeclick channel l load str store global phone number l load str store global wallet l load str store global isvisit l load str store global isjoin l load str store global isbotms l load str store global isloop l load str store global balance l build map store global set l load str store global coin l load const store name index l load const store global num l load str sit store global stat l load const none store global captcha l load const none store global sit l load str store global message l load str store global channel name l load const none store global channel id l load str store global bot name l load str store global coment l load const store name apver l build list store global bots l build list store global chns l build list store global grps l build list store global usrs l load const store global limit l load str store global lice l load str store global mase l load const store global ver l load const store global mass id l load const none store global click data l load str store global vipmes l load const store global jnum l load const store global ern l load const store global wiwnum l load const store global wiwcon l load const store name lops l build map store global setin l build list store global plogins l build map store global mesegs l build map store global admeseg l load str yes store global add l load const none store global minwd l load str numbers txt store global path l load const none store global proxy l load code load str isnumber make function neither defaults keyword only args annotations nor closures store name isnumber l load code load str request make function neither defaults keyword only args annotations nor closures store name request l load code load str getip make function neither defaults keyword only args annotations nor closures store name getip l load code load str send contact make function neither defaults keyword only args annotations nor closures store name send contact l load code load str send messege to members make function neither defaults keyword only args annotations nor closures store name send messege to members l load code load str chnum make function neither defaults keyword only args annotations nor closures store name chnum l load code load str numbers make function neither defaults keyword only args annotations nor closures store name numbers l load code load str anonsrobot make function neither defaults keyword only args annotations nor closures store name anonsrobot l load code load str start plogins make function neither defaults keyword only args annotations nor closures store name start plogins l load code load str send plogin messege make function neither defaults keyword only args annotations nor closures store name send plogin messege l load code load str send group messege make function neither defaults keyword only args annotations nor closures store name send group messege l load code load str settings make function neither defaults keyword only args annotations nor closures store name settings l load code load str send to privait group make function neither defaults keyword only args annotations nor closures store name send to privait group l load code load str creat user make function neither defaults keyword only args annotations nor closures store name creat user l load code load str add plogin make function neither defaults keyword only args annotations nor closures store name add plogin l load code load str delete plogin make function neither defaults keyword only args annotations nor closures store name delete plogin l load code load str getloop make function neither defaults keyword only args annotations nor closures store name getloop l load code load str add channel make function neither defaults keyword only args annotations nor closures store name add channel l load str no store global iswid l load code load str delete cannels make function neither defaults keyword only args annotations nor closures store name delete cannels l load code load str reset make function neither defaults keyword only args annotations nor closures store name reset l load code load str maseg make function neither defaults keyword only args annotations nor closures store name maseg l load code load str wait make function neither defaults keyword only args annotations nor closures store name wait l load code load str floodtimer make function neither defaults keyword only args annotations nor closures store name floodtimer l load code load str sleeptime make function neither defaults keyword only args annotations nor closures store name sleeptime l load code load str jointimer make function neither defaults keyword only args annotations nor closures store name jointimer l load code load str get username make function neither defaults keyword only args annotations nor closures store name get username l load code load str forward messages make function neither defaults keyword only args annotations nor closures store name forward messages l load code load str start make function neither defaults keyword only args annotations nor closures store name start l load code load str main make function neither defaults keyword only args annotations nor closures store name main come from come from come from l setup finally to l load name asyncio load method get event loop call method load method run until complete load name numbers call function call method pop top l load name lops load const inplace add store name lops l load global num get iter come from for iter to store name i l load name index load const inplace add store name index l load const store global jnum l load name asyncio load method get event loop call method load method run until complete load name main load name i call function call method pop top l load name sleep load const call function pop top jump back to come from l load name os load method system load name os load attr name load str nt compare op pop jump if false to load str cls jump forward to come from load str clear come from call method pop top l load name format load global ern load str call function store name er l load name format load global wiwcon load str call function store name wc l load global lice load str vip compare op pop jump if false to l load global lice format value load str load global limit format value load str build string store name vi jump forward to come from l load name fore load attr red format value load global lice format value load str load global limit format value load str build string store name vi come from l load name print load name banner call function pop top l load name print load name fore load attr green format value load str script license load name fore load attr red format value load str load name fore load attr reset format value load str load name vi format value build string call function pop top l load name print load name fore load attr green format value load str total loops load name fore load attr red format value load str load name fore load attr reset format value load str load name lops format value build string call function pop top l load name print load name fore load attr green format value load str total accounts load name fore load attr red format value load str load name fore load attr reset format value load str load name len load global num call function format value build string call function pop top l load name print load name fore load attr green format value load str accounts withdraw load name fore load attr red format value load str load name fore load attr reset format value load str load global wiwnum format value load str account build string call function pop top l load name print load name fore load attr green format value load str total withdraw load name fore load attr red format value load str load name fore load attr reset format value load str load name wc format value load str load global coin format value build string call function pop top l load name print load name fore load attr green format value load str total balance load name fore load attr red format value load str load name fore load attr reset format value load str load name er format value load str load global coin format value build string call function pop top l load name print load str n call function pop top l load global set load str autoloop binary subscr load method upper call method load str yes compare op pop jump if false to l pop block break loop to come from l load name sleeptime call function pop top l load const store global ern l load const store name index pop block jump back to come from finally l dup top load name keyboardinterrupt pop top pop top pop top l load name print load str call function pop top l load name sys load method exit call method pop top pop except jump back to jump back to come from parse error at or near dup top instruction at offset i also attached pyc file decompiled results and pydisasm results
| 0
|
6,211
| 9,124,270,989
|
IssuesEvent
|
2019-02-24 01:15:30
|
ahsankhatri/firestore-php
|
https://api.github.com/repos/ahsankhatri/firestore-php
|
closed
|
Error: PERMISSION_DENIED: Missing or insufficient permissions
|
enhancement in-process
|
I tried to write to FireStore but the server responded 403 PERMISSION_DENIED: Missing or insufficient permissions.
The database rule is:
```
service cloud.firestore {
match /databases/{database}/documents {
match /{document=**} {
allow read, write: if request.auth.uid != null;
}
}
}
```
If I change rule to `allow read, write: if true;` (turn off database security), it's OK.
|
1.0
|
Error: PERMISSION_DENIED: Missing or insufficient permissions - I tried to write to FireStore but the server responded 403 PERMISSION_DENIED: Missing or insufficient permissions.
The database rule is:
```
service cloud.firestore {
match /databases/{database}/documents {
match /{document=**} {
allow read, write: if request.auth.uid != null;
}
}
}
```
If I change rule to `allow read, write: if true;` (turn off database security), it's OK.
|
process
|
error permission denied missing or insufficient permissions i tried to write to firestore but the server responded permission denied missing or insufficient permissions the database rule is service cloud firestore match databases database documents match document allow read write if request auth uid null if i change rule to allow read write if true turn off database security it s ok
| 1
|
1,548
| 4,155,379,708
|
IssuesEvent
|
2016-06-16 14:45:58
|
openvstorage/openvstorage-health-check
|
https://api.github.com/repos/openvstorage/openvstorage-health-check
|
closed
|
Vpool is not present, ignore the test on that node
|
priority_urgent process_duplicate type_bug
|
If no vpool is present on a compute, controller ignore the test for that specific node.
|
1.0
|
Vpool is not present, ignore the test on that node - If no vpool is present on a compute, controller ignore the test for that specific node.
|
process
|
vpool is not present ignore the test on that node if no vpool is present on a compute controller ignore the test for that specific node
| 1
|
9,714
| 25,103,837,955
|
IssuesEvent
|
2022-11-08 15:19:49
|
MicrosoftDocs/architecture-center
|
https://api.github.com/repos/MicrosoftDocs/architecture-center
|
closed
|
Invalid C# syntax in code sample
|
doc-bug assigned-to-author triaged architecture-center/svc example-scenario/subsvc Pri2
|
Hello,
It appears that the code sample for the referenced documentation contains some invalid C# syntax, and given the code, it's not clear how to check if the device needs an SMS wakeup.
`if (var == [DeviceNotConnected] && connTimeOut == 0)`
I'm not sure where "DeviceNotConnected" is coming from, and I don't believe the use of `var` here or later in the method is valid.
Full code sample:
`TimeSpan connTimeOut = FromSeconds(0); // Period to wait for device to connect.
TimeSpan funcTimeOut = FromSeconds(30); // Period to wait for method to execute.
while (true) {
// Send the command via direct method. Initially use a timeout of zero
// for the connection, which determines whether the device is connected to
// IoT Hub or needs an SMS wakeup sent to it.
var method = new CloudToDeviceMethod("RemoteCommand", funcTimeOut, connTimeOut);
methodInvocation1.SetPayloadJson(CommandPayload);
var response = await serviceClient.InvokeDeviceMethodAsync(deviceId, method);
if (var == [DeviceNotConnected] && connTimeOut == 0) {
// The device is not currently connected and needs an SMS wakeup. This
// device should wake up within a period of < 30 seconds. Send the wakeup
// and retry the method request with a 30 second timeout on waiting for
// the device to connect.
connTimeOut = FromSeconds(30); // Set a 30 second connection timeout.
SendAsyncSMSWakeUpToDevice(); // Send SMS wakeup through mobile gateway.
continue; // Retry with new connection timeout.
} else {
// The method either succeeded or failed.
ActOnMethodResult(var);
break;
}
}`
Thank you!
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7b2f7457-51f0-b9e0-04bd-dabbe3586b34
* Version Independent ID: 08b9268e-c8d1-5259-0f0d-3fc851f35dbe
* Content: [IoT application-to-device commands - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/iot/cloud-to-device)
* Content Source: [docs/example-scenario/iot/cloud-to-device.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/iot/cloud-to-device.yml)
* Service: **architecture-center**
* Sub-service: **example-scenario**
* GitHub Login: @wamachine
* Microsoft Alias: **pnp**
|
1.0
|
Invalid C# syntax in code sample -
Hello,
It appears that the code sample for the referenced documentation contains some invalid C# syntax, and given the code, it's not clear how to check if the device needs an SMS wakeup.
`if (var == [DeviceNotConnected] && connTimeOut == 0)`
I'm not sure where "DeviceNotConnected" is coming from, and I don't believe the use of `var` here or later in the method is valid.
Full code sample:
`TimeSpan connTimeOut = FromSeconds(0); // Period to wait for device to connect.
TimeSpan funcTimeOut = FromSeconds(30); // Period to wait for method to execute.
while (true) {
// Send the command via direct method. Initially use a timeout of zero
// for the connection, which determines whether the device is connected to
// IoT Hub or needs an SMS wakeup sent to it.
var method = new CloudToDeviceMethod("RemoteCommand", funcTimeOut, connTimeOut);
methodInvocation1.SetPayloadJson(CommandPayload);
var response = await serviceClient.InvokeDeviceMethodAsync(deviceId, method);
if (var == [DeviceNotConnected] && connTimeOut == 0) {
// The device is not currently connected and needs an SMS wakeup. This
// device should wake up within a period of < 30 seconds. Send the wakeup
// and retry the method request with a 30 second timeout on waiting for
// the device to connect.
connTimeOut = FromSeconds(30); // Set a 30 second connection timeout.
SendAsyncSMSWakeUpToDevice(); // Send SMS wakeup through mobile gateway.
continue; // Retry with new connection timeout.
} else {
// The method either succeeded or failed.
ActOnMethodResult(var);
break;
}
}`
Thank you!
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7b2f7457-51f0-b9e0-04bd-dabbe3586b34
* Version Independent ID: 08b9268e-c8d1-5259-0f0d-3fc851f35dbe
* Content: [IoT application-to-device commands - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/iot/cloud-to-device)
* Content Source: [docs/example-scenario/iot/cloud-to-device.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/iot/cloud-to-device.yml)
* Service: **architecture-center**
* Sub-service: **example-scenario**
* GitHub Login: @wamachine
* Microsoft Alias: **pnp**
|
non_process
|
invalid c syntax in code sample hello it appears that the code sample for the referenced documentation contains some invalid c syntax and given the code it s not clear how to check if the device needs an sms wakeup if var conntimeout i m not sure where devicenotconnected is coming from and i don t believe the use of var here or later in the method is valid full code sample timespan conntimeout fromseconds period to wait for device to connect timespan functimeout fromseconds period to wait for method to execute while true send the command via direct method initially use a timeout of zero for the connection which determines whether the device is connected to iot hub or needs an sms wakeup sent to it var method new cloudtodevicemethod remotecommand functimeout conntimeout setpayloadjson commandpayload var response await serviceclient invokedevicemethodasync deviceid method if var conntimeout the device is not currently connected and needs an sms wakeup this device should wake up within a period of seconds send the wakeup and retry the method request with a second timeout on waiting for the device to connect conntimeout fromseconds set a second connection timeout sendasyncsmswakeuptodevice send sms wakeup through mobile gateway continue retry with new connection timeout else the method either succeeded or failed actonmethodresult var break thank you document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service example scenario github login wamachine microsoft alias pnp
| 0
|
39,335
| 12,663,382,544
|
IssuesEvent
|
2020-06-18 01:10:40
|
kenferrara/atlasdb
|
https://api.github.com/repos/kenferrara/atlasdb
|
opened
|
CVE-2020-7238 (High) detected in netty-codec-http-4.1.39.Final.jar, netty-all-4.1.39.Final.jar
|
security vulnerability
|
## CVE-2020-7238 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>netty-codec-http-4.1.39.Final.jar</b>, <b>netty-all-4.1.39.Final.jar</b></p></summary>
<p>
<details><summary><b>netty-codec-http-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- netty-handler-proxy-4.1.39.Final.jar (Root Library)
- :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library)
</details>
<details><summary><b>netty-all-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra-integration-tests/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- cassandra-all-2.2.13-1.6.0.jar (Root Library)
- :x: **netty-all-4.1.39.Final.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.
<p>Publish Date: 2020-01-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238>CVE-2020-7238</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/issues/9861">https://github.com/netty/netty/issues/9861</a></p>
<p>Release Date: 2020-01-27</p>
<p>Fix Resolution: io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"io.netty:netty-handler-proxy:4.1.39.Final;io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"},{"packageType":"Java","groupId":"io.netty","packageName":"netty-all","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"com.palantir.cassandra:cassandra-all:2.2.13-1.6.0;io.netty:netty-all:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"}],"vulnerabilityIdentifier":"CVE-2020-7238","vulnerabilityDetails":"Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-7238 (High) detected in netty-codec-http-4.1.39.Final.jar, netty-all-4.1.39.Final.jar - ## CVE-2020-7238 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>netty-codec-http-4.1.39.Final.jar</b>, <b>netty-all-4.1.39.Final.jar</b></p></summary>
<p>
<details><summary><b>netty-codec-http-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- netty-handler-proxy-4.1.39.Final.jar (Root Library)
- :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library)
</details>
<details><summary><b>netty-all-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra-integration-tests/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-all/4.1.39.Final/8de94e67d766103ac211e42f6cd9bb4b32e33e60/netty-all-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- cassandra-all-2.2.13-1.6.0.jar (Root Library)
- :x: **netty-all-4.1.39.Final.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.
<p>Publish Date: 2020-01-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238>CVE-2020-7238</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/issues/9861">https://github.com/netty/netty/issues/9861</a></p>
<p>Release Date: 2020-01-27</p>
<p>Fix Resolution: io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"io.netty:netty-handler-proxy:4.1.39.Final;io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"},{"packageType":"Java","groupId":"io.netty","packageName":"netty-all","packageVersion":"4.1.39.Final","isTransitiveDependency":true,"dependencyTree":"com.palantir.cassandra:cassandra-all:2.2.13-1.6.0;io.netty:netty-all:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.44.Final;io.netty:netty-codec-http:4.1.44.Final"}],"vulnerabilityIdentifier":"CVE-2020-7238","vulnerabilityDetails":"Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7238","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in netty codec http final jar netty all final jar cve high severity vulnerability vulnerable libraries netty codec http final jar netty all final jar netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file tmp ws scm atlasdb atlasdb cassandra build gradle path to vulnerable library root gradle caches modules files io netty netty codec http final netty codec http final jar root gradle caches modules files io netty netty codec http final netty codec http final jar dependency hierarchy netty handler proxy final jar root library x netty codec http final jar vulnerable library netty all final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file tmp ws scm atlasdb atlasdb cassandra integration tests build gradle path to vulnerable library root gradle caches modules files io netty netty all final netty all final jar root gradle caches modules files io netty netty all final netty all final jar dependency hierarchy cassandra all jar root library x netty all final jar vulnerable library vulnerability details netty final allows http request smuggling because it mishandles transfer encoding whitespace such as a transfer encoding chunked line and a later content length header this issue exists because of an incomplete fix for cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all final io netty netty codec http final isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails netty final allows http request smuggling because it mishandles transfer encoding whitespace such as a transfer encoding chunked line and a later content length header this issue exists because of an incomplete fix for cve vulnerabilityurl
| 0
|
207,389
| 7,127,842,932
|
IssuesEvent
|
2018-01-21 03:16:40
|
DASSL/ClassDB
|
https://api.github.com/repos/DASSL/ClassDB
|
closed
|
The docs-development process may not support adding image and other files (WM)
|
missing priority low wrong
|
**PS:** I am creating this issue so we can collectively discuss and receive notifications as new comments are added.
Context: We have add at least two instances where non-markdown files are to be included in the docs folder. See #136 and #139.
In general, we have the need to add HTML/Visio/image files to the docs folder with design notes and these files will have to be linked to from docs pages which are markdown (MD) files.
On the surface, this does not sound like much of an issue, but I worry our development process for docs may not support this ability: at least not very easily. Our docs-development process uses the Wiki as the "development environment" to write the docs and it worked very well in M1: We need to use the Wiki so anyone can easily update any page directly online and also avoid the approval process which can be both labor and time intensive.
It seems, if we want to continue to use the Wiki for doc development, we will need to clone the Wiki as a repository, add files to it, and check in the file. Even then, I am not sure if it will work correctly as evidenced in [@afig's Wiki page](https://github.com/DASSL/ClassDB/wiki/createRole()-User-Schema-checks-and-actions) on the flowchart related to Issue #136. See how the embedded image referencing a remote URL does not show up.
We need to **urgently** understand the underlying issues and learn workarounds/solutions so documentation does not become a problem.
Perhaps there is no issue at all, in which case, it will be helpful to confirm so before we get too far and then find a problem.
|
1.0
|
The docs-development process may not support adding image and other files (WM) - **PS:** I am creating this issue so we can collectively discuss and receive notifications as new comments are added.
Context: We have add at least two instances where non-markdown files are to be included in the docs folder. See #136 and #139.
In general, we have the need to add HTML/Visio/image files to the docs folder with design notes and these files will have to be linked to from docs pages which are markdown (MD) files.
On the surface, this does not sound like much of an issue, but I worry our development process for docs may not support this ability: at least not very easily. Our docs-development process uses the Wiki as the "development environment" to write the docs and it worked very well in M1: We need to use the Wiki so anyone can easily update any page directly online and also avoid the approval process which can be both labor and time intensive.
It seems, if we want to continue to use the Wiki for doc development, we will need to clone the Wiki as a repository, add files to it, and check in the file. Even then, I am not sure if it will work correctly as evidenced in [@afig's Wiki page](https://github.com/DASSL/ClassDB/wiki/createRole()-User-Schema-checks-and-actions) on the flowchart related to Issue #136. See how the embedded image referencing a remote URL does not show up.
We need to **urgently** understand the underlying issues and learn workarounds/solutions so documentation does not become a problem.
Perhaps there is no issue at all, in which case, it will be helpful to confirm so before we get too far and then find a problem.
|
non_process
|
the docs development process may not support adding image and other files wm ps i am creating this issue so we can collectively discuss and receive notifications as new comments are added context we have add at least two instances where non markdown files are to be included in the docs folder see and in general we have the need to add html visio image files to the docs folder with design notes and these files will have to be linked to from docs pages which are markdown md files on the surface this does not sound like much of an issue but i worry our development process for docs may not support this ability at least not very easily our docs development process uses the wiki as the development environment to write the docs and it worked very well in we need to use the wiki so anyone can easily update any page directly online and also avoid the approval process which can be both labor and time intensive it seems if we want to continue to use the wiki for doc development we will need to clone the wiki as a repository add files to it and check in the file even then i am not sure if it will work correctly as evidenced in on the flowchart related to issue see how the embedded image referencing a remote url does not show up we need to urgently understand the underlying issues and learn workarounds solutions so documentation does not become a problem perhaps there is no issue at all in which case it will be helpful to confirm so before we get too far and then find a problem
| 0
|
8,279
| 11,434,738,769
|
IssuesEvent
|
2020-02-04 17:58:25
|
prisma/prisma-client-js
|
https://api.github.com/repos/prisma/prisma-client-js
|
closed
|
Update graphics to use schema.prisma instead of datamodel.mdl
|
process/candidate
|
## What
Currently the graphics use `datamodel.mdl`, e.g.
<img width="751" alt="Screenshot 2020-02-04 at 15 51 58" src="https://user-images.githubusercontent.com/1992255/73755537-7269e400-4766-11ea-8bb7-420b4a5b0135.png">
They should be updated to `schema.prisma`
|
1.0
|
Update graphics to use schema.prisma instead of datamodel.mdl - ## What
Currently the graphics use `datamodel.mdl`, e.g.
<img width="751" alt="Screenshot 2020-02-04 at 15 51 58" src="https://user-images.githubusercontent.com/1992255/73755537-7269e400-4766-11ea-8bb7-420b4a5b0135.png">
They should be updated to `schema.prisma`
|
process
|
update graphics to use schema prisma instead of datamodel mdl what currently the graphics use datamodel mdl e g img width alt screenshot at src they should be updated to schema prisma
| 1
|
1,601
| 3,070,690,084
|
IssuesEvent
|
2015-08-19 07:06:20
|
NatLibFi/Skosmos
|
https://api.github.com/repos/NatLibFi/Skosmos
|
opened
|
Make use of Jena3 text index for better performance
|
branch enhancement performance size-large
|
There are significant changes (implemented by Alexis Miara and myself) in the [jena-text](http://jena.staging.apache.org/documentation/query/text-query.html) module of Jena 3.0.0 / Fuseki 1.3.0 / Fuseki 2.3.0. These include
* support for storing language tags of literals and limiting queries to a specific language
* support for storing full literal values in the index and accessing them at query time
* support for deleting obsolete entries from the text index
These together enable a new way of using the text index from Skosmos:
* Text queries could, in most cases, be limited to a specific language. This avoids false hits from the text index that would have to be filtered out using SPARQL, and should thus speed up queries, particularly the alphabetical display for large vocabularies.
* Since the text index can return full literal values, there is less need to find out which literal value actually matched the query (using regular expressions or string matching functions, as is done currently). This should make text index related SPARQL queries both simpler and faster.
* The `uidField` should be enabled, so that stale entries will be dropped from the text index. Currently the performance of text index related queries deteriorates slightly each time the vocabulary data is updated. This is probably due to stale entries. Cleaning them up should prevent this performance deterioration.
Text index related code in JenaTextSparql (and possibly GenericSparql) will need to be heavily rewritten. Luckily the new code should be simpler than the old one and we already have pretty good unit tests for this functionality, so it is easy to verify what works and what doesn't.
Text index configuration needs to be changed to enable the new features, and text indexes must then be rebuilt. Fuseki 1.3.0/2.3.0 will require Java 8 to be installed on servers, development machines and the Travis CI environment (where it should be available, but not used by default).
(Finto project note: this is a way of implementing FINTO-85: Tuki hyvin suurille tietovarannoille)
|
True
|
Make use of Jena3 text index for better performance - There are significant changes (implemented by Alexis Miara and myself) in the [jena-text](http://jena.staging.apache.org/documentation/query/text-query.html) module of Jena 3.0.0 / Fuseki 1.3.0 / Fuseki 2.3.0. These include
* support for storing language tags of literals and limiting queries to a specific language
* support for storing full literal values in the index and accessing them at query time
* support for deleting obsolete entries from the text index
These together enable a new way of using the text index from Skosmos:
* Text queries could, in most cases, be limited to a specific language. This avoids false hits from the text index that would have to be filtered out using SPARQL, and should thus speed up queries, particularly the alphabetical display for large vocabularies.
* Since the text index can return full literal values, there is less need to find out which literal value actually matched the query (using regular expressions or string matching functions, as is done currently). This should make text index related SPARQL queries both simpler and faster.
* The `uidField` should be enabled, so that stale entries will be dropped from the text index. Currently the performance of text index related queries deteriorates slightly each time the vocabulary data is updated. This is probably due to stale entries. Cleaning them up should prevent this performance deterioration.
Text index related code in JenaTextSparql (and possibly GenericSparql) will need to be heavily rewritten. Luckily the new code should be simpler than the old one and we already have pretty good unit tests for this functionality, so it is easy to verify what works and what doesn't.
Text index configuration needs to be changed to enable the new features, and text indexes must then be rebuilt. Fuseki 1.3.0/2.3.0 will require Java 8 to be installed on servers, development machines and the Travis CI environment (where it should be available, but not used by default).
(Finto project note: this is a way of implementing FINTO-85: Tuki hyvin suurille tietovarannoille)
|
non_process
|
make use of text index for better performance there are significant changes implemented by alexis miara and myself in the module of jena fuseki fuseki these include support for storing language tags of literals and limiting queries to a specific language support for storing full literal values in the index and accessing them at query time support for deleting obsolete entries from the text index these together enable a new way of using the text index from skosmos text queries could in most cases be limited to a specific language this avoids false hits from the text index that would have to be filtered out using sparql and should thus speed up queries particularly the alphabetical display for large vocabularies since the text index can return full literal values there is less need to find out which literal value actually matched the query using regular expressions or string matching functions as is done currently this should make text index related sparql queries both simpler and faster the uidfield should be enabled so that stale entries will be dropped from the text index currently the performance of text index related queries deteriorates slightly each time the vocabulary data is updated this is probably due to stale entries cleaning them up should prevent this performance deterioration text index related code in jenatextsparql and possibly genericsparql will need to be heavily rewritten luckily the new code should be simpler than the old one and we already have pretty good unit tests for this functionality so it is easy to verify what works and what doesn t text index configuration needs to be changed to enable the new features and text indexes must then be rebuilt fuseki will require java to be installed on servers development machines and the travis ci environment where it should be available but not used by default finto project note this is a way of implementing finto tuki hyvin suurille tietovarannoille
| 0
|
12,127
| 2,685,016,856
|
IssuesEvent
|
2015-03-29 16:24:00
|
IssueMigrationTest/Test5
|
https://api.github.com/repos/IssueMigrationTest/Test5
|
closed
|
*ERROR* basedados.py:11: cannot locate module: psycopg2
|
auto-migrated Priority-Medium Type-Defect
|
**Issue by luisl...@gmail.com**
_9 Jul 2009 at 2:09 GMT_
_Originally opened on Google Code_
----
```
Hi,
I have a module (basedados.py) that starts as:
"""
#-*- coding: iso-8859-15 -*-
import ConfigParser
import psycopg2
"""
It runs fine in Python, as stand-alone or when called from other modules.
If I try:
$ shedskin basedados
*** SHED SKIN Python-to-C++ Compiler 0.1.1 ***
Copyright 2005-2009 Mark Dufour; License GNU GPL version 3 (See LICENSE)
*ERROR* basedados.py:11: cannot locate module: psycopg2
Even if I remove every line from the module except the 'import psycopg2'
statement, the error is the same.
I'm using Slackware Linux 12.1, python 2.5.2 and shedskin 0.1.1
```
|
1.0
|
*ERROR* basedados.py:11: cannot locate module: psycopg2 - **Issue by luisl...@gmail.com**
_9 Jul 2009 at 2:09 GMT_
_Originally opened on Google Code_
----
```
Hi,
I have a module (basedados.py) that starts as:
"""
#-*- coding: iso-8859-15 -*-
import ConfigParser
import psycopg2
"""
It runs fine in Python, as stand-alone or when called from other modules.
If I try:
$ shedskin basedados
*** SHED SKIN Python-to-C++ Compiler 0.1.1 ***
Copyright 2005-2009 Mark Dufour; License GNU GPL version 3 (See LICENSE)
*ERROR* basedados.py:11: cannot locate module: psycopg2
Even if I remove every line from the module except the 'import psycopg2'
statement, the error is the same.
I'm using Slackware Linux 12.1, python 2.5.2 and shedskin 0.1.1
```
|
non_process
|
error basedados py cannot locate module issue by luisl gmail com jul at gmt originally opened on google code hi i have a module basedados py that starts as coding iso import configparser import it runs fine in python as stand alone or when called from other modules if i try shedskin basedados shed skin python to c compiler copyright mark dufour license gnu gpl version see license error basedados py cannot locate module even if i remove every line from the module except the import statement the error is the same i m using slackware linux python and shedskin
| 0
|
14,936
| 18,365,441,937
|
IssuesEvent
|
2021-10-10 00:48:00
|
FTBTeam/FTB-App-Feedback
|
https://api.github.com/repos/FTBTeam/FTB-App-Feedback
|
closed
|
[Bug] Some files still are stored in default install location (/users/my_username/.ftba) even when install location is changed
|
bug bug:subprocess effect:linux bug:unconfirmed subprocess
|
**Operating System:**
OS: Windows
**Installer Version:**
Installer:
**App Version:**
App: 202007161516-479dedc615
**UI Version:**
UI: 39f37b8d
**Describe the bug**
Even when I change in install location in the launcher, some necessary files are installed in the default location (in the bin folder). It seems these files are used for the vanilla launcher. If I delete the folder, the launcher no longer can run instances. If I move the files from there to my chosen install location, the launcher still no longer can run instances.
**To Reproduce**
Steps to reproduce the behaviour:
1. Change install location
2. create a new instance
3. run the instance once
4. Observe existence of .ftba folder in home directory with necessary files still located there
**Expected behaviour**
All files necessary to the launcher are installed in the chosen install location
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
Add any other context about the problem here.
|
2.0
|
[Bug] Some files still are stored in default install location (/users/my_username/.ftba) even when install location is changed - **Operating System:**
OS: Windows
**Installer Version:**
Installer:
**App Version:**
App: 202007161516-479dedc615
**UI Version:**
UI: 39f37b8d
**Describe the bug**
Even when I change in install location in the launcher, some necessary files are installed in the default location (in the bin folder). It seems these files are used for the vanilla launcher. If I delete the folder, the launcher no longer can run instances. If I move the files from there to my chosen install location, the launcher still no longer can run instances.
**To Reproduce**
Steps to reproduce the behaviour:
1. Change install location
2. create a new instance
3. run the instance once
4. Observe existence of .ftba folder in home directory with necessary files still located there
**Expected behaviour**
All files necessary to the launcher are installed in the chosen install location
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
Add any other context about the problem here.
|
process
|
some files still are stored in default install location users my username ftba even when install location is changed operating system os windows installer version installer app version app ui version ui describe the bug even when i change in install location in the launcher some necessary files are installed in the default location in the bin folder it seems these files are used for the vanilla launcher if i delete the folder the launcher no longer can run instances if i move the files from there to my chosen install location the launcher still no longer can run instances to reproduce steps to reproduce the behaviour change install location create a new instance run the instance once observe existence of ftba folder in home directory with necessary files still located there expected behaviour all files necessary to the launcher are installed in the chosen install location screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here
| 1
|
10,473
| 12,421,425,777
|
IssuesEvent
|
2020-05-23 16:48:11
|
facebook/hhvm
|
https://api.github.com/repos/facebook/hhvm
|
closed
|
Unable append to array returned by function as referece
|
php5 incompatibility probably easy
|
In this case (http://3v4l.org/DmV7V) appending to array fails (unlike PHP interpreter)
``` php
<?php
class Test {
private $array = [];
public function __construct() {
$i = 0;
while ($i < 10) {
$this->array[$i] = str_repeat(" ", 1e6);
$i++;
}
}
public function &variables() {
return $this->array;
}
}
$test = new Test();
$test->variables()[10] = str_repeat(" ", 1e6); // ok
$test->variables()[] = str_repeat(" ", 1e6); // error, why ?
echo count($test->variables()) . "\n";
```
|
True
|
Unable append to array returned by function as referece - In this case (http://3v4l.org/DmV7V) appending to array fails (unlike PHP interpreter)
``` php
<?php
class Test {
private $array = [];
public function __construct() {
$i = 0;
while ($i < 10) {
$this->array[$i] = str_repeat(" ", 1e6);
$i++;
}
}
public function &variables() {
return $this->array;
}
}
$test = new Test();
$test->variables()[10] = str_repeat(" ", 1e6); // ok
$test->variables()[] = str_repeat(" ", 1e6); // error, why ?
echo count($test->variables()) . "\n";
```
|
non_process
|
unable append to array returned by function as referece in this case appending to array fails unlike php interpreter php php class test private array public function construct i while i this array str repeat i public function variables return this array test new test test variables str repeat ok test variables str repeat error why echo count test variables n
| 0
|
500,555
| 14,502,050,319
|
IssuesEvent
|
2020-12-11 20:25:38
|
GetSpeckle/speckle-browser-extension
|
https://api.github.com/repos/GetSpeckle/speckle-browser-extension
|
closed
|
Upgrade polkadot libs to support edgeware nodes
|
Priority: Medium
|
While speckle is still using `@polkadot/api 0.76.1` and `@polkadot/keyring 0.76.1`, `@polkadot/api 0.81.1` and `@polkadot/keyring 0.93.1` are already released as of writing.
`edgeware-node-types` relies on those new libs, hence speckle cannot support edgeware until we upgrade to those versions. Certain interfaces are changed in those versions, `keyring-vault` needs to be updated accordingly
|
1.0
|
Upgrade polkadot libs to support edgeware nodes - While speckle is still using `@polkadot/api 0.76.1` and `@polkadot/keyring 0.76.1`, `@polkadot/api 0.81.1` and `@polkadot/keyring 0.93.1` are already released as of writing.
`edgeware-node-types` relies on those new libs, hence speckle cannot support edgeware until we upgrade to those versions. Certain interfaces are changed in those versions, `keyring-vault` needs to be updated accordingly
|
non_process
|
upgrade polkadot libs to support edgeware nodes while speckle is still using polkadot api and polkadot keyring polkadot api and polkadot keyring are already released as of writing edgeware node types relies on those new libs hence speckle cannot support edgeware until we upgrade to those versions certain interfaces are changed in those versions keyring vault needs to be updated accordingly
| 0
|
13,911
| 16,668,775,354
|
IssuesEvent
|
2021-06-07 08:19:52
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
opened
|
Release 4.2 - June 2021
|
P1 release team-XProduct type: process
|
# Status of Bazel 4.2
This release will use Bazel 4.1.0 as its baseline and we will apply selected cherry-picks and backports on top of it. Please request cherry-picks that you'd like to get into Bazel 4.2.0 here via a comment.
- Expected release date: June 2021
- [List of release blockers](https://github.com/bazelbuild/bazel/labels/Release%20blocker)
To report a release-blocking bug, please file a bug using the `Release blocker` label, and cc me.
Task list:
- [ ] Pick release baseline: https://github.com/bazelbuild/bazel/commit/c4975efdd7ede7b46637bf353209d9ac371181a5
- [ ] Create release candidate: https://releases.bazel.build/4.2.0/rc1/
- [ ] Check downstream projects: ...
- [ ] [Create draft release announcement](https://docs.google.com/document/d/1wDvulLlj4NAlPZamdlEVFORks3YXJonCjyuQMUQEmB0/edit)
- [ ] Send for review the release announcement PR: ...
- [ ] Push the release, notify package maintainers: ...
- [ ] Update the documentation
- [ ] Push the blog post
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/tag/4.2.0)
|
1.0
|
Release 4.2 - June 2021 - # Status of Bazel 4.2
This release will use Bazel 4.1.0 as its baseline and we will apply selected cherry-picks and backports on top of it. Please request cherry-picks that you'd like to get into Bazel 4.2.0 here via a comment.
- Expected release date: June 2021
- [List of release blockers](https://github.com/bazelbuild/bazel/labels/Release%20blocker)
To report a release-blocking bug, please file a bug using the `Release blocker` label, and cc me.
Task list:
- [ ] Pick release baseline: https://github.com/bazelbuild/bazel/commit/c4975efdd7ede7b46637bf353209d9ac371181a5
- [ ] Create release candidate: https://releases.bazel.build/4.2.0/rc1/
- [ ] Check downstream projects: ...
- [ ] [Create draft release announcement](https://docs.google.com/document/d/1wDvulLlj4NAlPZamdlEVFORks3YXJonCjyuQMUQEmB0/edit)
- [ ] Send for review the release announcement PR: ...
- [ ] Push the release, notify package maintainers: ...
- [ ] Update the documentation
- [ ] Push the blog post
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/tag/4.2.0)
|
process
|
release june status of bazel this release will use bazel as its baseline and we will apply selected cherry picks and backports on top of it please request cherry picks that you d like to get into bazel here via a comment expected release date june to report a release blocking bug please file a bug using the release blocker label and cc me task list pick release baseline create release candidate check downstream projects send for review the release announcement pr push the release notify package maintainers update the documentation push the blog post update the
| 1
|
17,729
| 23,636,906,817
|
IssuesEvent
|
2022-08-25 13:58:01
|
argosp/trialdash
|
https://api.github.com/repos/argosp/trialdash
|
closed
|
Position in one shape entities from different types
|
Priority High in process
|
in the current staged app there is no option to position in one shape entities from different types.
right now as demand we want to enable this option. its involved with the change of logic of existing implementation.
for now the develop continues with the current option. waiting for clarity with this issue.
|
1.0
|
Position in one shape entities from different types - in the current staged app there is no option to position in one shape entities from different types.
right now as demand we want to enable this option. its involved with the change of logic of existing implementation.
for now the develop continues with the current option. waiting for clarity with this issue.
|
process
|
position in one shape entities from different types in the current staged app there is no option to position in one shape entities from different types right now as demand we want to enable this option its involved with the change of logic of existing implementation for now the develop continues with the current option waiting for clarity with this issue
| 1
|
236,478
| 19,547,835,280
|
IssuesEvent
|
2022-01-02 07:16:50
|
Misaka-0x447f/iis.cloudy-misaka
|
https://api.github.com/repos/Misaka-0x447f/iis.cloudy-misaka
|
opened
|
Set Up This Blog - Internet Information Service @ misaka.org
|
Gitalk https://iis.misaka.org/blog/test-post2/
|
https://iis.misaka.org/blog/test-post2/
This is a guide for everything to set up in the repository to make this place your own! Regardless of if you're new to Gridsome or just looking for the checklist of things to update when you clone this project, this post is for you!
|
1.0
|
Set Up This Blog - Internet Information Service @ misaka.org - https://iis.misaka.org/blog/test-post2/
This is a guide for everything to set up in the repository to make this place your own! Regardless of if you're new to Gridsome or just looking for the checklist of things to update when you clone this project, this post is for you!
|
non_process
|
set up this blog internet information service misaka org this is a guide for everything to set up in the repository to make this place your own regardless of if you re new to gridsome or just looking for the checklist of things to update when you clone this project this post is for you
| 0
|
100,305
| 11,185,035,697
|
IssuesEvent
|
2019-12-31 21:52:35
|
packethost/docs
|
https://api.github.com/repos/packethost/docs
|
closed
|
local-and-global-bgp.md
|
product-documentation
|
[local-and-global-bgp.md](https://github.com/packethost/docs/blob/master/products/04-network/03-advanced/02-local-and-global-bgp.md)
This file is too big for its current purpose. Please work on to take out some part of it to our /guides section and link it here. For more info - @jacobsmith928
|
1.0
|
local-and-global-bgp.md - [local-and-global-bgp.md](https://github.com/packethost/docs/blob/master/products/04-network/03-advanced/02-local-and-global-bgp.md)
This file is too big for its current purpose. Please work on to take out some part of it to our /guides section and link it here. For more info - @jacobsmith928
|
non_process
|
local and global bgp md this file is too big for its current purpose please work on to take out some part of it to our guides section and link it here for more info
| 0
|
274,589
| 23,851,930,901
|
IssuesEvent
|
2022-09-06 18:48:26
|
lowRISC/opentitan
|
https://api.github.com/repos/lowRISC/opentitan
|
opened
|
[chip-test] chip_sw_pwrmgr_main_power_glitch_reset extension
|
Component:ChipLevelTest
|
### Test point name
[chip_sw_pwrmgr_main_power_glitch_reset](https://github.com/lowRISC/opentitan/blob/620c8c0c07dc0ab05448c0e1ad707d82072f58da/hw/top_earlgrey/data/chip_testplan.hjson#L1035)
### Host side component
_No response_
### OpenTitanTool infrastructure implemented
_No response_
### Contact person
@matutem, @tjaychen
### Checklist
Add SVAs in order to check that:
- Check that the clock valids are deasserted if a power glitch is detected.
- clocks are stopped if their valid is deasserted
The assertions in the second bulled point above can probably go into RTL.
Please fill out this checklist as items are completed. Link to PRs and issues as appropriate.
- [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes)
- [ ] Device-side (C) component developed
- [ ] Bazel build rules developed
- [ ] Host-side component developed
- [ ] HJSON test plan updated with test name (so it shows up in the dashboard)
- [ ] Test added to dvsim nightly regression (and passing at time of checking)
|
1.0
|
[chip-test] chip_sw_pwrmgr_main_power_glitch_reset extension - ### Test point name
[chip_sw_pwrmgr_main_power_glitch_reset](https://github.com/lowRISC/opentitan/blob/620c8c0c07dc0ab05448c0e1ad707d82072f58da/hw/top_earlgrey/data/chip_testplan.hjson#L1035)
### Host side component
_No response_
### OpenTitanTool infrastructure implemented
_No response_
### Contact person
@matutem, @tjaychen
### Checklist
Add SVAs in order to check that:
- Check that the clock valids are deasserted if a power glitch is detected.
- clocks are stopped if their valid is deasserted
The assertions in the second bulled point above can probably go into RTL.
Please fill out this checklist as items are completed. Link to PRs and issues as appropriate.
- [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes)
- [ ] Device-side (C) component developed
- [ ] Bazel build rules developed
- [ ] Host-side component developed
- [ ] HJSON test plan updated with test name (so it shows up in the dashboard)
- [ ] Test added to dvsim nightly regression (and passing at time of checking)
|
non_process
|
chip sw pwrmgr main power glitch reset extension test point name host side component no response opentitantool infrastructure implemented no response contact person matutem tjaychen checklist add svas in order to check that check that the clock valids are deasserted if a power glitch is detected clocks are stopped if their valid is deasserted the assertions in the second bulled point above can probably go into rtl please fill out this checklist as items are completed link to prs and issues as appropriate check if existing test covers most or all of this testpoint if so either extend said test to cover all points or skip the next checkboxes device side c component developed bazel build rules developed host side component developed hjson test plan updated with test name so it shows up in the dashboard test added to dvsim nightly regression and passing at time of checking
| 0
|
11,116
| 13,957,682,306
|
IssuesEvent
|
2020-10-24 08:07:40
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
PL: A kind request to initiate the harvesting process
|
Geoportal Harvesting process PL - Poland
|
Hi Helpdesk Team,
can we kindly ask you to push harvesting process for PL catalogue?
Regards,
Piotr Perz
|
1.0
|
PL: A kind request to initiate the harvesting process - Hi Helpdesk Team,
can we kindly ask you to push harvesting process for PL catalogue?
Regards,
Piotr Perz
|
process
|
pl a kind request to initiate the harvesting process hi helpdesk team can we kindly ask you to push harvesting process for pl catalogue regards piotr perz
| 1
|
21,645
| 30,079,567,391
|
IssuesEvent
|
2023-06-29 01:12:37
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
@pandacss/dev 0.5.0 has 1 guarddog issues
|
npm-silent-process-execution
|
```{"npm-silent-process-execution":[{"code":" (0, import_node_child_process.spawn)(import_node_process7.default.execPath, [import_node_path3.default.join(__dirname2, \"check.js\"), JSON.stringify(this.#options)], {\n detached: true,\n stdio: \"ignore\"\n }).unref();","location":"package/dist/cli-main.js:17813","message":"This package is silently executing another executable"}]}```
|
1.0
|
@pandacss/dev 0.5.0 has 1 guarddog issues - ```{"npm-silent-process-execution":[{"code":" (0, import_node_child_process.spawn)(import_node_process7.default.execPath, [import_node_path3.default.join(__dirname2, \"check.js\"), JSON.stringify(this.#options)], {\n detached: true,\n stdio: \"ignore\"\n }).unref();","location":"package/dist/cli-main.js:17813","message":"This package is silently executing another executable"}]}```
|
process
|
pandacss dev has guarddog issues npm silent process execution n detached true n stdio ignore n unref location package dist cli main js message this package is silently executing another executable
| 1
|
234,290
| 19,123,540,987
|
IssuesEvent
|
2021-12-01 03:00:41
|
libreswan/libreswan
|
https://api.github.com/repos/libreswan/libreswan
|
closed
|
f35: addconn needs more seccomp system calls
|
regression testing
|
```
+type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX
+type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX
```
```
+#0 __GI___fstatat64 (fd=7, file=0x7f18d8b5fff5 "", buf=0x7ffc76ee2110, flag=4096) at ../sysdeps/unix/sysv/linux/fstatat64.c:162
+#1 0x00007f18d8a1cb43 in __GI__IO_file_doallocate (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947
+#2 0x00007f18d8a2a890 in __GI__IO_doallocbuf (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947
+#3 __GI__IO_doallocbuf (fp=fp@entry=0x7f18d7d66e20) at genops.c:342
+#4 0x00007f18d8a299bc in _IO_new_file_underflow (fp=0x7f18d7d66e20) at fileops.c:486
+#5 0x00007f18d8a2a946 in __GI__IO_default_uflow (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947
+#6 0x00007f18d8a1e27c in __GI__IO_getline_info (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1, eof=eof@entry=0x0) at iogetline.c:60
+#7 0x00007f18d8a1e37c in __GI__IO_getline (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1) at iogetline.c:34
+#8 0x00007f18d8a1d360 in _IO_fgets (buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1024, fp=fp@entry=0x7f18d7d66e20) at iofgets.c:53
+#9 0x00007f18d8c5fb73 in fgets (__stream=0x7f18d7d66e20, __n=1024, __s=0x7ffc76ee2310 "\317\036") at PATH/include/bits/stdio2.h:272
+#10 ub_ctx_hosts (ctx=0x7f18d8066e70, fname=fname@entry=0x55ebeee4fbbe "/etc/hosts") at libunbound/libunbound.c:1215
+#11 0x000055ebeee3e02a in unbound_ctx_config (do_dnssec=<optimized out>, rootfile=0x7f18d8058fe0 "/var/lib/unbound/root.key", trusted=0x0, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:63
+#12 0x000055ebeee3e6bb in unbound_sync_init (do_dnssec=<optimized out>, rootfile=<optimized out>, trusted=<optimized out>, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:226
+#13 0x000055ebeee21b27 in main (argc=2, argv=0x7ffc76ee3258) at /source/programs/addconn/addconn.c:416
```
|
1.0
|
f35: addconn needs more seccomp system calls - ```
+type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX
+type=SECCOMP msg=audit(XXX): auid=AUID uid=0 gid=0 ses=SES subj=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 pid=PID comm="addconn" exe="PATH/libexec/ipsec/addconn" sig=31 arch=c000003e syscall=262 compat=0 ip=XXX
```
```
+#0 __GI___fstatat64 (fd=7, file=0x7f18d8b5fff5 "", buf=0x7ffc76ee2110, flag=4096) at ../sysdeps/unix/sysv/linux/fstatat64.c:162
+#1 0x00007f18d8a1cb43 in __GI__IO_file_doallocate (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947
+#2 0x00007f18d8a2a890 in __GI__IO_doallocbuf (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947
+#3 __GI__IO_doallocbuf (fp=fp@entry=0x7f18d7d66e20) at genops.c:342
+#4 0x00007f18d8a299bc in _IO_new_file_underflow (fp=0x7f18d7d66e20) at fileops.c:486
+#5 0x00007f18d8a2a946 in __GI__IO_default_uflow (fp=0x7f18d7d66e20) at PATH/src/debug/glibc-2.34-8.fc35.x86_64/libio/libioP.h:947
+#6 0x00007f18d8a1e27c in __GI__IO_getline_info (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1, eof=eof@entry=0x0) at iogetline.c:60
+#7 0x00007f18d8a1e37c in __GI__IO_getline (fp=fp@entry=0x7f18d7d66e20, buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1023, delim=delim@entry=10, extract_delim=extract_delim@entry=1) at iogetline.c:34
+#8 0x00007f18d8a1d360 in _IO_fgets (buf=buf@entry=0x7ffc76ee2310 "\317\036", n=n@entry=1024, fp=fp@entry=0x7f18d7d66e20) at iofgets.c:53
+#9 0x00007f18d8c5fb73 in fgets (__stream=0x7f18d7d66e20, __n=1024, __s=0x7ffc76ee2310 "\317\036") at PATH/include/bits/stdio2.h:272
+#10 ub_ctx_hosts (ctx=0x7f18d8066e70, fname=fname@entry=0x55ebeee4fbbe "/etc/hosts") at libunbound/libunbound.c:1215
+#11 0x000055ebeee3e02a in unbound_ctx_config (do_dnssec=<optimized out>, rootfile=0x7f18d8058fe0 "/var/lib/unbound/root.key", trusted=0x0, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:63
+#12 0x000055ebeee3e6bb in unbound_sync_init (do_dnssec=<optimized out>, rootfile=<optimized out>, trusted=<optimized out>, logger=logger@entry=0x55ebeee64520 <progname_logger>) at /source/lib/libswan/unbound.c:226
+#13 0x000055ebeee21b27 in main (argc=2, argv=0x7ffc76ee3258) at /source/programs/addconn/addconn.c:416
```
|
non_process
|
addconn needs more seccomp system calls type seccomp msg audit xxx auid auid uid gid ses ses subj unconfined u unconfined r unconfined t pid pid comm addconn exe path libexec ipsec addconn sig arch syscall compat ip xxx type seccomp msg audit xxx auid auid uid gid ses ses subj unconfined u unconfined r unconfined t pid pid comm addconn exe path libexec ipsec addconn sig arch syscall compat ip xxx gi fd file buf flag at sysdeps unix sysv linux c in gi io file doallocate fp at path src debug glibc libio libiop h in gi io doallocbuf fp at path src debug glibc libio libiop h gi io doallocbuf fp fp entry at genops c in io new file underflow fp at fileops c in gi io default uflow fp at path src debug glibc libio libiop h in gi io getline info fp fp entry buf buf entry n n entry delim delim entry extract delim extract delim entry eof eof entry at iogetline c in gi io getline fp fp entry buf buf entry n n entry delim delim entry extract delim extract delim entry at iogetline c in io fgets buf buf entry n n entry fp fp entry at iofgets c in fgets stream n s at path include bits h ub ctx hosts ctx fname fname entry etc hosts at libunbound libunbound c in unbound ctx config do dnssec rootfile var lib unbound root key trusted logger logger entry at source lib libswan unbound c in unbound sync init do dnssec rootfile trusted logger logger entry at source lib libswan unbound c in main argc argv at source programs addconn addconn c
| 0
|
437,921
| 12,604,653,678
|
IssuesEvent
|
2020-06-11 15:17:07
|
eobanb/indianapublicmedia-web
|
https://api.github.com/repos/eobanb/indianapublicmedia-web
|
closed
|
WYSIWYG editor - custom formatting
|
enhancement medium priority
|
(note: as described [in Hannon Hill's documentation](https://www.hannonhill.com/cascadecms/latest/content-authoring/wysiwyg-editor-configurations/index.html))
This warrants testing / developing a proof of concept for several possible uses:
- /ethergame/ hints that can be hidden/shown individually
- /news/ 'view older updates' button that can be applied to a post without HTML knowledge
- inline Google DFP ads (I.e. for /indiandroid/ and /theinbox/) without copying/pasting HTML blocks
|
1.0
|
WYSIWYG editor - custom formatting - (note: as described [in Hannon Hill's documentation](https://www.hannonhill.com/cascadecms/latest/content-authoring/wysiwyg-editor-configurations/index.html))
This warrants testing / developing a proof of concept for several possible uses:
- /ethergame/ hints that can be hidden/shown individually
- /news/ 'view older updates' button that can be applied to a post without HTML knowledge
- inline Google DFP ads (I.e. for /indiandroid/ and /theinbox/) without copying/pasting HTML blocks
|
non_process
|
wysiwyg editor custom formatting note as described this warrants testing developing a proof of concept for several possible uses ethergame hints that can be hidden shown individually news view older updates button that can be applied to a post without html knowledge inline google dfp ads i e for indiandroid and theinbox without copying pasting html blocks
| 0
|
15,099
| 18,836,033,166
|
IssuesEvent
|
2021-11-11 01:04:19
|
GoogleCloudPlatform/cloud-ops-sandbox
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
|
opened
|
Ensure dependency bot is tracking terraform dependencies
|
type: process priority: p3
|
[We recently had an issue where a terraform dependency updated and broke our released code](https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/pull/864). I believe this repo should have a bot to catch these issues and auto-update dependencies (dependabot, rennovatebot, or both), but we should confirm this
|
1.0
|
Ensure dependency bot is tracking terraform dependencies - [We recently had an issue where a terraform dependency updated and broke our released code](https://github.com/GoogleCloudPlatform/cloud-ops-sandbox/pull/864). I believe this repo should have a bot to catch these issues and auto-update dependencies (dependabot, rennovatebot, or both), but we should confirm this
|
process
|
ensure dependency bot is tracking terraform dependencies i believe this repo should have a bot to catch these issues and auto update dependencies dependabot rennovatebot or both but we should confirm this
| 1
|
11,309
| 14,112,001,026
|
IssuesEvent
|
2020-11-07 02:44:15
|
e4exp/paper_manager_abstract
|
https://api.github.com/repos/e4exp/paper_manager_abstract
|
opened
|
Learning Music Helps You Read: Using Transfer to Study Linguistic Structure in Language Models
|
2020 Natural Language Processing Programming Languages Transfer Learning _read_later
|
* https://arxiv.org/abs/2004.14601
* 2020
本研究では、ニューラル言語モデルにおける文法構造の符号化を解析するための手法として、伝達学習を提案する。
本研究では、LSTMを非言語データ上で訓練し、自然言語上での性能を評価することで、どのようなデータがLSTMが自然言語で使用できる一般化可能な構造特徴を誘発するかを評価する。
その結果、非言語データに潜在的な構造を持つデータ(MIDI音楽やJavaコード)を学習させると、表面形式や語彙に重なりがないにもかかわらず、自然言語でのテスト性能が向上することがわかった。
この改善につながるモデルがどのような抽象構造をコード化しているのかを特定するために、我々は2つの人工括弧言語を用いて同様の実験を行った:
1つは階層的な再帰構造を持つ言語、もう1つはペアトークンを持つが再帰を持たない制御言語である。
驚くべきことに、これらの人工言語のいずれかでモデルを訓練すると、自然言語でテストしたときに同じような実質的な利益が得られる。
さらに、語彙の重複を制御した自然言語間の移動に関する実験では、テスト言語でのゼロショット性能は、訓練言語との類型論的な統語的類似性と高い相関があることが示された。
この結果は、ニューラルモデルが抽象的な統語構造を表現する方法や、自然言語の習得を可能にする構造的帰納的バイアスの種類についての洞察を提供する。
|
1.0
|
Learning Music Helps You Read: Using Transfer to Study Linguistic Structure in Language Models - * https://arxiv.org/abs/2004.14601
* 2020
本研究では、ニューラル言語モデルにおける文法構造の符号化を解析するための手法として、伝達学習を提案する。
本研究では、LSTMを非言語データ上で訓練し、自然言語上での性能を評価することで、どのようなデータがLSTMが自然言語で使用できる一般化可能な構造特徴を誘発するかを評価する。
その結果、非言語データに潜在的な構造を持つデータ(MIDI音楽やJavaコード)を学習させると、表面形式や語彙に重なりがないにもかかわらず、自然言語でのテスト性能が向上することがわかった。
この改善につながるモデルがどのような抽象構造をコード化しているのかを特定するために、我々は2つの人工括弧言語を用いて同様の実験を行った:
1つは階層的な再帰構造を持つ言語、もう1つはペアトークンを持つが再帰を持たない制御言語である。
驚くべきことに、これらの人工言語のいずれかでモデルを訓練すると、自然言語でテストしたときに同じような実質的な利益が得られる。
さらに、語彙の重複を制御した自然言語間の移動に関する実験では、テスト言語でのゼロショット性能は、訓練言語との類型論的な統語的類似性と高い相関があることが示された。
この結果は、ニューラルモデルが抽象的な統語構造を表現する方法や、自然言語の習得を可能にする構造的帰納的バイアスの種類についての洞察を提供する。
|
process
|
learning music helps you read using transfer to study linguistic structure in language models 本研究では、ニューラル言語モデルにおける文法構造の符号化を解析するための手法として、伝達学習を提案する。 本研究では、lstmを非言語データ上で訓練し、自然言語上での性能を評価することで、どのようなデータがlstmが自然言語で使用できる一般化可能な構造特徴を誘発するかを評価する。 その結果、非言語データに潜在的な構造を持つデータ midi音楽やjavaコード を学習させると、表面形式や語彙に重なりがないにもかかわらず、自然言語でのテスト性能が向上することがわかった。 この改善につながるモデルがどのような抽象構造をコード化しているのかを特定するために、 : 、 。 驚くべきことに、これらの人工言語のいずれかでモデルを訓練すると、自然言語でテストしたときに同じような実質的な利益が得られる。 さらに、語彙の重複を制御した自然言語間の移動に関する実験では、テスト言語でのゼロショット性能は、訓練言語との類型論的な統語的類似性と高い相関があることが示された。 この結果は、ニューラルモデルが抽象的な統語構造を表現する方法や、自然言語の習得を可能にする構造的帰納的バイアスの種類についての洞察を提供する。
| 1
|
274,622
| 23,853,210,457
|
IssuesEvent
|
2022-09-06 20:04:23
|
mountaincharlie/project-five-midi-dragon
|
https://api.github.com/repos/mountaincharlie/project-five-midi-dragon
|
closed
|
Testing
|
Testing
|
- [x] check the landing page renders correctly
- [x] check that all styling has been properly applied
- [x] create test_urls.py to test home app url
- [x] create test_views.py to test home app view
- [x] document tests in README
|
1.0
|
Testing - - [x] check the landing page renders correctly
- [x] check that all styling has been properly applied
- [x] create test_urls.py to test home app url
- [x] create test_views.py to test home app view
- [x] document tests in README
|
non_process
|
testing check the landing page renders correctly check that all styling has been properly applied create test urls py to test home app url create test views py to test home app view document tests in readme
| 0
|
138,383
| 18,785,304,236
|
IssuesEvent
|
2021-11-08 11:28:18
|
sultanabubaker/gradle-simple-project
|
https://api.github.com/repos/sultanabubaker/gradle-simple-project
|
reopened
|
CVE-2020-10673 (High) detected in jackson-databind-2.8.11.6.jar
|
security vulnerability
|
## CVE-2020-10673 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: gradle-simple-project/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.6/35753201d0cdb1dbe998ab289bca1180b68d4368/jackson-databind-2.8.11.6.jar</p>
<p>
Dependency Hierarchy:
- sami-pr-nexus-2.0.0.jar (Root Library)
- core-5.0.0.jar
- crypto-5.0.0.jar
- :x: **jackson-databind-2.8.11.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/gradle-simple-project/commit/591721fb6f22aef1255695db7f0486533bff325a">591721fb6f22aef1255695db7f0486533bff325a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.6","packageFilePaths":["/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.jitpack:sami-pr-nexus:2.0.0;org.web3j:core:5.0.0;org.web3j:crypto:5.0.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-10673 (High) detected in jackson-databind-2.8.11.6.jar - ## CVE-2020-10673 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: gradle-simple-project/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.11.6/35753201d0cdb1dbe998ab289bca1180b68d4368/jackson-databind-2.8.11.6.jar</p>
<p>
Dependency Hierarchy:
- sami-pr-nexus-2.0.0.jar (Root Library)
- core-5.0.0.jar
- crypto-5.0.0.jar
- :x: **jackson-databind-2.8.11.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/gradle-simple-project/commit/591721fb6f22aef1255695db7f0486533bff325a">591721fb6f22aef1255695db7f0486533bff325a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.6","packageFilePaths":["/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.jitpack:sami-pr-nexus:2.0.0;org.web3j:core:5.0.0;org.web3j:crypto:5.0.0;com.fasterxml.jackson.core:jackson-databind:2.8.11.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-10673","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file gradle simple project build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy sami pr nexus jar root library core jar crypto jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io jitpack sami pr nexus org core org crypto com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus vulnerabilityurl
| 0
|
60,995
| 14,939,098,501
|
IssuesEvent
|
2021-01-25 16:32:27
|
EIDSS/EIDSS7
|
https://api.github.com/repos/EIDSS/EIDSS7
|
closed
|
LAB17 Approve samples to be destroyed
|
Build 98.0 Minor bug
|
**Summary**
Step 9. Verification step: in the same section enter the Lab Sample ID in the search field and select search.The approved samples records display in the Samples tab. The approved samples records display with <Sample Status> Deleted.
**To Reproduce**
Steps to reproduce the behavior:
1. Log in as lisasimpson
2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search.
3. The approved samples records display in the Samples tab.
4. The approved samples records display with <Sample Status> Deleted.
**Expected behavior**
2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search.
3. The approved samples records display in the Samples tab.
4. The approved samples records display with <Sample Status> Destroyed.
**Screenshots**

**Additional details:**
- Build:
- Script title (enter ad hoc if not script-based):
**Issue severity (Optional)**
Severity (critical, major, minor, low):
**Additional context**
Add any other context about the problem here.
|
1.0
|
LAB17 Approve samples to be destroyed - **Summary**
Step 9. Verification step: in the same section enter the Lab Sample ID in the search field and select search.The approved samples records display in the Samples tab. The approved samples records display with <Sample Status> Deleted.
**To Reproduce**
Steps to reproduce the behavior:
1. Log in as lisasimpson
2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search.
3. The approved samples records display in the Samples tab.
4. The approved samples records display with <Sample Status> Deleted.
**Expected behavior**
2. Go to Step 9: Verification step: in the same section enter the Lab Sample ID in the search field and select search.
3. The approved samples records display in the Samples tab.
4. The approved samples records display with <Sample Status> Destroyed.
**Screenshots**

**Additional details:**
- Build:
- Script title (enter ad hoc if not script-based):
**Issue severity (Optional)**
Severity (critical, major, minor, low):
**Additional context**
Add any other context about the problem here.
|
non_process
|
approve samples to be destroyed summary step verification step in the same section enter the lab sample id in the search field and select search the approved samples records display in the samples tab the approved samples records display with deleted to reproduce steps to reproduce the behavior log in as lisasimpson go to step verification step in the same section enter the lab sample id in the search field and select search the approved samples records display in the samples tab the approved samples records display with deleted expected behavior go to step verification step in the same section enter the lab sample id in the search field and select search the approved samples records display in the samples tab the approved samples records display with destroyed screenshots additional details build script title enter ad hoc if not script based issue severity optional severity critical major minor low additional context add any other context about the problem here
| 0
|
4,875
| 7,753,598,812
|
IssuesEvent
|
2018-05-31 01:38:07
|
kubernetes-sigs/kubebuilder
|
https://api.github.com/repos/kubernetes-sigs/kubebuilder
|
closed
|
Create a SECURITY_CONTACTS file.
|
process
|
As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS
file.
The template for the file can be found in the kubernetes-template repository[2].
A description for the file is in the steering-committee docs[3], you might need
to search that page for "Security Contacts".
Please feel free to ping me on the PR when you make it, otherwise I will see when
you close this issue. :)
Thanks so much, let me know if you have any questions.
(This issue was generated from a tool, apologies for any weirdness.)
[1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE
[2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS
[3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
|
1.0
|
Create a SECURITY_CONTACTS file. - As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS
file.
The template for the file can be found in the kubernetes-template repository[2].
A description for the file is in the steering-committee docs[3], you might need
to search that page for "Security Contacts".
Please feel free to ping me on the PR when you make it, otherwise I will see when
you close this issue. :)
Thanks so much, let me know if you have any questions.
(This issue was generated from a tool, apologies for any weirdness.)
[1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE
[2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS
[3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
|
process
|
create a security contacts file as per the email sent to kubernetes dev please create a security contacts file the template for the file can be found in the kubernetes template repository a description for the file is in the steering committee docs you might need to search that page for security contacts please feel free to ping me on the pr when you make it otherwise i will see when you close this issue thanks so much let me know if you have any questions this issue was generated from a tool apologies for any weirdness
| 1
|
106,148
| 4,263,799,489
|
IssuesEvent
|
2016-07-12 02:45:34
|
ampproject/docs
|
https://api.github.com/repos/ampproject/docs
|
closed
|
Invalid word breaking
|
Priority: High
|
Word breaking is invalid, unacceptable. The culprit must be hiding somewhere [in these lines](https://github.com/ampproject/docs/blob/5ce07f344ad482d23e34032e052d3849861cd6ba/_sass/content.scss#L42).

|
1.0
|
Invalid word breaking - Word breaking is invalid, unacceptable. The culprit must be hiding somewhere [in these lines](https://github.com/ampproject/docs/blob/5ce07f344ad482d23e34032e052d3849861cd6ba/_sass/content.scss#L42).

|
non_process
|
invalid word breaking word breaking is invalid unacceptable the culprit must be hiding somewhere
| 0
|
1,827
| 4,613,585,929
|
IssuesEvent
|
2016-09-25 03:25:34
|
EBrown8534/StackExchangeStatisticsExplorer
|
https://api.github.com/repos/EBrown8534/StackExchangeStatisticsExplorer
|
closed
|
Going to `/Sites/Compare/` with no `Sites` Query String parameter throws error
|
bug in process
|
Self explanatory, if you go to the `/Sites/Compare/` endpoint without a valid `Sites` query string parameter it should redirect to `/Sites/Select/` just like `/Sites/Detail/` redirects to `/Sites/Default/` without a valid `Site` query string parameter.
|
1.0
|
Going to `/Sites/Compare/` with no `Sites` Query String parameter throws error - Self explanatory, if you go to the `/Sites/Compare/` endpoint without a valid `Sites` query string parameter it should redirect to `/Sites/Select/` just like `/Sites/Detail/` redirects to `/Sites/Default/` without a valid `Site` query string parameter.
|
process
|
going to sites compare with no sites query string parameter throws error self explanatory if you go to the sites compare endpoint without a valid sites query string parameter it should redirect to sites select just like sites detail redirects to sites default without a valid site query string parameter
| 1
|
4,856
| 7,745,641,995
|
IssuesEvent
|
2018-05-29 18:59:54
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
The invoke does not work.
|
cxp in-process storage/svc triaged
|
Invoke-StorageSyncFileRecall : The term 'Invoke-StorageSyncFileRecall' is not recognized as the name of a cmdlet. This is despite having ran the Import-Module.
Also, on the Import-Module command for StorageSync.Management.ServerCmdlets.dll complains about unapproved verbs.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 4814031f-3cca-10a6-8587-cf1a045d2481
* Version Independent ID: b96d1378-9a11-c00e-1c84-a69d4ff1cff3
* Content: [Manage registered servers with Azure File Sync (preview)](https://docs.microsoft.com/en-us/azure/storage/files/storage-sync-files-server-registration#feedback)
* Content Source: [articles/storage/files/storage-sync-files-server-registration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/storage/files/storage-sync-files-server-registration.md)
* Service: **storage**
* Product: **unspecified**
* GitHub Login: @wmgries
* Microsoft Alias: **wgries**
|
1.0
|
The invoke does not work. - Invoke-StorageSyncFileRecall : The term 'Invoke-StorageSyncFileRecall' is not recognized as the name of a cmdlet. This is despite having ran the Import-Module.
Also, on the Import-Module command for StorageSync.Management.ServerCmdlets.dll complains about unapproved verbs.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 4814031f-3cca-10a6-8587-cf1a045d2481
* Version Independent ID: b96d1378-9a11-c00e-1c84-a69d4ff1cff3
* Content: [Manage registered servers with Azure File Sync (preview)](https://docs.microsoft.com/en-us/azure/storage/files/storage-sync-files-server-registration#feedback)
* Content Source: [articles/storage/files/storage-sync-files-server-registration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/storage/files/storage-sync-files-server-registration.md)
* Service: **storage**
* Product: **unspecified**
* GitHub Login: @wmgries
* Microsoft Alias: **wgries**
|
process
|
the invoke does not work invoke storagesyncfilerecall the term invoke storagesyncfilerecall is not recognized as the name of a cmdlet this is despite having ran the import module also on the import module command for storagesync management servercmdlets dll complains about unapproved verbs document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service storage product unspecified github login wmgries microsoft alias wgries
| 1
|
21,590
| 29,983,918,107
|
IssuesEvent
|
2023-06-25 02:00:07
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Fri, 23 Jun 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### Exploring the Role of Audio in Video Captioning
- **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS)
- **Arxiv link:** https://arxiv.org/abs/2306.12559
- **Pdf link:** https://arxiv.org/pdf/2306.12559
- **Abstract**
Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Data-Free Backbone Fine-Tuning for Pruned Neural Networks
- **Authors:** Adrian Holzbock, Achyut Hegde, Klaus Dietmayer, Vasileios Belagiannis
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.12881
- **Pdf link:** https://arxiv.org/pdf/2306.12881
- **Abstract**
Model compression techniques reduce the computational load and memory consumption of deep neural networks. After the compression operation, e.g. parameter pruning, the model is normally fine-tuned on the original training dataset to recover from the performance drop caused by compression. However, the training data is not always available due to privacy issues or other factors. In this work, we present a data-free fine-tuning approach for pruning the backbone of deep neural networks. In particular, the pruned network backbone is trained with synthetically generated images, and our proposed intermediate supervision to mimic the unpruned backbone's output feature map. Afterwards, the pruned backbone can be combined with the original network head to make predictions. We generate synthetic images by back-propagating gradients to noise images while relying on L1-pruning for the backbone pruning. In our experiments, we show that our approach is task-independent due to pruning only the backbone. By evaluating our approach on 2D human pose estimation, object detection, and image classification, we demonstrate promising performance compared to the unpruned model. Our code is available at https://github.com/holzbock/dfbf.
## Keyword: RAW
### Exploring the Role of Audio in Video Captioning
- **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS)
- **Arxiv link:** https://arxiv.org/abs/2306.12559
- **Pdf link:** https://arxiv.org/pdf/2306.12559
- **Abstract**
Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input.
### Neural Spectro-polarimetric Fields
- **Authors:** Youngchan Kim, Wonjoon Jin, Sunghyun Cho, Seung-Hwan Baek
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2306.12562
- **Pdf link:** https://arxiv.org/pdf/2306.12562
- **Abstract**
Modeling the spatial radiance distribution of light rays in a scene has been extensively explored for applications, including view synthesis. Spectrum and polarization, the wave properties of light, are often neglected due to their integration into three RGB spectral bands and their non-perceptibility to human vision. Despite this, these properties encompass substantial material and geometric information about a scene. In this work, we propose to model spectro-polarimetric fields, the spatial Stokes-vector distribution of any light ray at an arbitrary wavelength. We present Neural Spectro-polarimetric Fields (NeSpoF), a neural representation that models the physically-valid Stokes vector at given continuous variables of position, direction, and wavelength. NeSpoF manages inherently noisy raw measurements, showcases memory efficiency, and preserves physically vital signals, factors that are crucial for representing the high-dimensional signal of a spectro-polarimetric field. To validate NeSpoF, we introduce the first multi-view hyperspectral-polarimetric image dataset, comprised of both synthetic and real-world scenes. These were captured using our compact hyperspectral-polarimetric imaging system, which has been calibrated for robustness against system imperfections. We demonstrate the capabilities of NeSpoF on diverse scenes.
### DreamEdit: Subject-driven Image Editing
- **Authors:** Tianle Li, Max Ku, Cong Wei, Wenhu Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.12624
- **Pdf link:** https://arxiv.org/pdf/2306.12624
- **Abstract**
Subject-driven image generation aims at generating images containing customized subjects, which has recently drawn enormous attention from the research community. However, the previous works cannot precisely control the background and position of the target subject. In this work, we aspire to fill the void and propose two novel subject-driven sub-tasks, i.e., Subject Replacement and Subject Addition. The new tasks are challenging in multiple aspects: replacing a subject with a customized one can change its shape, texture, and color, while adding a target subject to a designated position in a provided scene necessitates a context-aware posture. To conquer these two novel tasks, we first manually curate a new dataset DreamEditBench containing 22 different types of subjects, and 440 source images with different difficulty levels. We plan to host DreamEditBench as a platform and hire trained evaluators for standard human evaluation. We also devise an innovative method DreamEditor to resolve these tasks by performing iterative generation, which enables a smooth adaptation to the customized subject. In this project, we conduct automatic and human evaluations to understand the performance of DreamEditor and baselines on DreamEditBench. For Subject Replacement, we found that the existing models are sensitive to the shape and color of the original subject. The model failure rate will dramatically increase when the source and target subjects are highly different. For Subject Addition, we found that the existing models cannot easily blend the customized subjects into the background smoothly, leading to noticeable artifacts in the generated image. We hope DreamEditBench can become a standard platform to enable future investigations toward building more controllable subject-driven image editing. Our project homepage is https://dreameditbenchteam.github.io/.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Fri, 23 Jun 23 - ## Keyword: events
### Exploring the Role of Audio in Video Captioning
- **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS)
- **Arxiv link:** https://arxiv.org/abs/2306.12559
- **Pdf link:** https://arxiv.org/pdf/2306.12559
- **Abstract**
Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
There is no result
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Data-Free Backbone Fine-Tuning for Pruned Neural Networks
- **Authors:** Adrian Holzbock, Achyut Hegde, Klaus Dietmayer, Vasileios Belagiannis
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.12881
- **Pdf link:** https://arxiv.org/pdf/2306.12881
- **Abstract**
Model compression techniques reduce the computational load and memory consumption of deep neural networks. After the compression operation, e.g. parameter pruning, the model is normally fine-tuned on the original training dataset to recover from the performance drop caused by compression. However, the training data is not always available due to privacy issues or other factors. In this work, we present a data-free fine-tuning approach for pruning the backbone of deep neural networks. In particular, the pruned network backbone is trained with synthetically generated images, and our proposed intermediate supervision to mimic the unpruned backbone's output feature map. Afterwards, the pruned backbone can be combined with the original network head to make predictions. We generate synthetic images by back-propagating gradients to noise images while relying on L1-pruning for the backbone pruning. In our experiments, we show that our approach is task-independent due to pruning only the backbone. By evaluating our approach on 2D human pose estimation, object detection, and image classification, we demonstrate promising performance compared to the unpruned model. Our code is available at https://github.com/holzbock/dfbf.
## Keyword: RAW
### Exploring the Role of Audio in Video Captioning
- **Authors:** Yuhan Shen, Linjie Yang, Longyin Wen, Haichao Yu, Ehsan Elhamifar, Heng Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Sound (cs.SD); Audio and Speech Processing (eess.AS)
- **Arxiv link:** https://arxiv.org/abs/2306.12559
- **Pdf link:** https://arxiv.org/pdf/2306.12559
- **Abstract**
Recent focus in video captioning has been on designing architectures that can consume both video and text modalities, and using large-scale video datasets with text transcripts for pre-training, such as HowTo100M. Though these approaches have achieved significant improvement, the audio modality is often ignored in video captioning. In this work, we present an audio-visual framework, which aims to fully exploit the potential of the audio modality for captioning. Instead of relying on text transcripts extracted via automatic speech recognition (ASR), we argue that learning with raw audio signals can be more beneficial, as audio has additional information including acoustic events, speaker identity, etc. Our contributions are twofold. First, we observed that the model overspecializes to the audio modality when pre-training with both video and audio modality, since the ground truth (i.e., text transcripts) can be solely predicted using audio. We proposed a Modality Balanced Pre-training (MBP) loss to mitigate this issue and significantly improve the performance on downstream tasks. Second, we slice and dice different design choices of the cross-modal module, which may become an information bottleneck and generate inferior results. We proposed new local-global fusion mechanisms to improve information exchange across audio and video. We demonstrate significant improvements by leveraging the audio modality on four datasets, and even outperform the state of the art on some metrics without relying on the text modality as the input.
### Neural Spectro-polarimetric Fields
- **Authors:** Youngchan Kim, Wonjoon Jin, Sunghyun Cho, Seung-Hwan Baek
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2306.12562
- **Pdf link:** https://arxiv.org/pdf/2306.12562
- **Abstract**
Modeling the spatial radiance distribution of light rays in a scene has been extensively explored for applications, including view synthesis. Spectrum and polarization, the wave properties of light, are often neglected due to their integration into three RGB spectral bands and their non-perceptibility to human vision. Despite this, these properties encompass substantial material and geometric information about a scene. In this work, we propose to model spectro-polarimetric fields, the spatial Stokes-vector distribution of any light ray at an arbitrary wavelength. We present Neural Spectro-polarimetric Fields (NeSpoF), a neural representation that models the physically-valid Stokes vector at given continuous variables of position, direction, and wavelength. NeSpoF manages inherently noisy raw measurements, showcases memory efficiency, and preserves physically vital signals, factors that are crucial for representing the high-dimensional signal of a spectro-polarimetric field. To validate NeSpoF, we introduce the first multi-view hyperspectral-polarimetric image dataset, comprised of both synthetic and real-world scenes. These were captured using our compact hyperspectral-polarimetric imaging system, which has been calibrated for robustness against system imperfections. We demonstrate the capabilities of NeSpoF on diverse scenes.
### DreamEdit: Subject-driven Image Editing
- **Authors:** Tianle Li, Max Ku, Cong Wei, Wenhu Chen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2306.12624
- **Pdf link:** https://arxiv.org/pdf/2306.12624
- **Abstract**
Subject-driven image generation aims at generating images containing customized subjects, which has recently drawn enormous attention from the research community. However, the previous works cannot precisely control the background and position of the target subject. In this work, we aspire to fill the void and propose two novel subject-driven sub-tasks, i.e., Subject Replacement and Subject Addition. The new tasks are challenging in multiple aspects: replacing a subject with a customized one can change its shape, texture, and color, while adding a target subject to a designated position in a provided scene necessitates a context-aware posture. To conquer these two novel tasks, we first manually curate a new dataset DreamEditBench containing 22 different types of subjects, and 440 source images with different difficulty levels. We plan to host DreamEditBench as a platform and hire trained evaluators for standard human evaluation. We also devise an innovative method DreamEditor to resolve these tasks by performing iterative generation, which enables a smooth adaptation to the customized subject. In this project, we conduct automatic and human evaluations to understand the performance of DreamEditor and baselines on DreamEditBench. For Subject Replacement, we found that the existing models are sensitive to the shape and color of the original subject. The model failure rate will dramatically increase when the source and target subjects are highly different. For Subject Addition, we found that the existing models cannot easily blend the customized subjects into the background smoothly, leading to noticeable artifacts in the generated image. We hope DreamEditBench can become a standard platform to enable future investigations toward building more controllable subject-driven image editing. Our project homepage is https://dreameditbenchteam.github.io/.
## Keyword: raw image
There is no result
|
process
|
new submissions for fri jun keyword events exploring the role of audio in video captioning authors yuhan shen linjie yang longyin wen haichao yu ehsan elhamifar heng wang subjects computer vision and pattern recognition cs cv sound cs sd audio and speech processing eess as arxiv link pdf link abstract recent focus in video captioning has been on designing architectures that can consume both video and text modalities and using large scale video datasets with text transcripts for pre training such as though these approaches have achieved significant improvement the audio modality is often ignored in video captioning in this work we present an audio visual framework which aims to fully exploit the potential of the audio modality for captioning instead of relying on text transcripts extracted via automatic speech recognition asr we argue that learning with raw audio signals can be more beneficial as audio has additional information including acoustic events speaker identity etc our contributions are twofold first we observed that the model overspecializes to the audio modality when pre training with both video and audio modality since the ground truth i e text transcripts can be solely predicted using audio we proposed a modality balanced pre training mbp loss to mitigate this issue and significantly improve the performance on downstream tasks second we slice and dice different design choices of the cross modal module which may become an information bottleneck and generate inferior results we proposed new local global fusion mechanisms to improve information exchange across audio and video we demonstrate significant improvements by leveraging the audio modality on four datasets and even outperform the state of the art on some metrics without relying on the text modality as the input keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression data free backbone fine tuning for pruned neural networks authors adrian holzbock achyut hegde klaus dietmayer vasileios belagiannis subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract model compression techniques reduce the computational load and memory consumption of deep neural networks after the compression operation e g parameter pruning the model is normally fine tuned on the original training dataset to recover from the performance drop caused by compression however the training data is not always available due to privacy issues or other factors in this work we present a data free fine tuning approach for pruning the backbone of deep neural networks in particular the pruned network backbone is trained with synthetically generated images and our proposed intermediate supervision to mimic the unpruned backbone s output feature map afterwards the pruned backbone can be combined with the original network head to make predictions we generate synthetic images by back propagating gradients to noise images while relying on pruning for the backbone pruning in our experiments we show that our approach is task independent due to pruning only the backbone by evaluating our approach on human pose estimation object detection and image classification we demonstrate promising performance compared to the unpruned model our code is available at keyword raw exploring the role of audio in video captioning authors yuhan shen linjie yang longyin wen haichao yu ehsan elhamifar heng wang subjects computer vision and pattern recognition cs cv sound cs sd audio and speech processing eess as arxiv link pdf link abstract recent focus in video captioning has been on designing architectures that can consume both video and text modalities and using large scale video datasets with text transcripts for pre training such as though these approaches have achieved significant improvement the audio modality is often ignored in video captioning in this work we present an audio visual framework which aims to fully exploit the potential of the audio modality for captioning instead of relying on text transcripts extracted via automatic speech recognition asr we argue that learning with raw audio signals can be more beneficial as audio has additional information including acoustic events speaker identity etc our contributions are twofold first we observed that the model overspecializes to the audio modality when pre training with both video and audio modality since the ground truth i e text transcripts can be solely predicted using audio we proposed a modality balanced pre training mbp loss to mitigate this issue and significantly improve the performance on downstream tasks second we slice and dice different design choices of the cross modal module which may become an information bottleneck and generate inferior results we proposed new local global fusion mechanisms to improve information exchange across audio and video we demonstrate significant improvements by leveraging the audio modality on four datasets and even outperform the state of the art on some metrics without relying on the text modality as the input neural spectro polarimetric fields authors youngchan kim wonjoon jin sunghyun cho seung hwan baek subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract modeling the spatial radiance distribution of light rays in a scene has been extensively explored for applications including view synthesis spectrum and polarization the wave properties of light are often neglected due to their integration into three rgb spectral bands and their non perceptibility to human vision despite this these properties encompass substantial material and geometric information about a scene in this work we propose to model spectro polarimetric fields the spatial stokes vector distribution of any light ray at an arbitrary wavelength we present neural spectro polarimetric fields nespof a neural representation that models the physically valid stokes vector at given continuous variables of position direction and wavelength nespof manages inherently noisy raw measurements showcases memory efficiency and preserves physically vital signals factors that are crucial for representing the high dimensional signal of a spectro polarimetric field to validate nespof we introduce the first multi view hyperspectral polarimetric image dataset comprised of both synthetic and real world scenes these were captured using our compact hyperspectral polarimetric imaging system which has been calibrated for robustness against system imperfections we demonstrate the capabilities of nespof on diverse scenes dreamedit subject driven image editing authors tianle li max ku cong wei wenhu chen subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract subject driven image generation aims at generating images containing customized subjects which has recently drawn enormous attention from the research community however the previous works cannot precisely control the background and position of the target subject in this work we aspire to fill the void and propose two novel subject driven sub tasks i e subject replacement and subject addition the new tasks are challenging in multiple aspects replacing a subject with a customized one can change its shape texture and color while adding a target subject to a designated position in a provided scene necessitates a context aware posture to conquer these two novel tasks we first manually curate a new dataset dreameditbench containing different types of subjects and source images with different difficulty levels we plan to host dreameditbench as a platform and hire trained evaluators for standard human evaluation we also devise an innovative method dreameditor to resolve these tasks by performing iterative generation which enables a smooth adaptation to the customized subject in this project we conduct automatic and human evaluations to understand the performance of dreameditor and baselines on dreameditbench for subject replacement we found that the existing models are sensitive to the shape and color of the original subject the model failure rate will dramatically increase when the source and target subjects are highly different for subject addition we found that the existing models cannot easily blend the customized subjects into the background smoothly leading to noticeable artifacts in the generated image we hope dreameditbench can become a standard platform to enable future investigations toward building more controllable subject driven image editing our project homepage is keyword raw image there is no result
| 1
|
17,871
| 5,529,950,527
|
IssuesEvent
|
2017-03-21 00:24:42
|
bcgov/api-specs
|
https://api.github.com/repos/bcgov/api-specs
|
closed
|
Auto generate localities where HWY 97 doesn't have street directions
|
data prep scripts enhancement GEOCODER medium priority
|
Graham to generate a spreadsheet of in the form required by the address prep scripts. Michael to review and send back to Graham who will then integrate the locality list into the address prep scripts.
|
1.0
|
Auto generate localities where HWY 97 doesn't have street directions - Graham to generate a spreadsheet of in the form required by the address prep scripts. Michael to review and send back to Graham who will then integrate the locality list into the address prep scripts.
|
non_process
|
auto generate localities where hwy doesn t have street directions graham to generate a spreadsheet of in the form required by the address prep scripts michael to review and send back to graham who will then integrate the locality list into the address prep scripts
| 0
|
118,060
| 25,243,222,077
|
IssuesEvent
|
2022-11-15 09:12:52
|
arduino/arduino-cli
|
https://api.github.com/repos/arduino/arduino-cli
|
closed
|
Boost libraries with the exact match
|
type: enhancement topic: code
|
### Describe the request
> When I search for SD I expect to see [SD](https://github.com/arduino-libraries/SD) to show up in the top of the list so I can install it without a lot of scrolling through the list of results. But instead it is far down in the results.
Related to https://github.com/arduino/arduino-ide/issues/1106
### Describe the current behavior
`arduino-cli lib search SD`
[out.txt](https://github.com/arduino/arduino-cli/files/10005220/out.txt)
As you can see the SD lib is pretty down in the list
### Arduino CLI version
0.28.0
### Operating system
N/A
### Operating system version
na
### Additional context
_No response_
### Issue checklist
- [X] I searched for previous requests in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=)
- [X] I verified the feature was still missing when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds)
- [X] My request contains all necessary details
|
1.0
|
Boost libraries with the exact match - ### Describe the request
> When I search for SD I expect to see [SD](https://github.com/arduino-libraries/SD) to show up in the top of the list so I can install it without a lot of scrolling through the list of results. But instead it is far down in the results.
Related to https://github.com/arduino/arduino-ide/issues/1106
### Describe the current behavior
`arduino-cli lib search SD`
[out.txt](https://github.com/arduino/arduino-cli/files/10005220/out.txt)
As you can see the SD lib is pretty down in the list
### Arduino CLI version
0.28.0
### Operating system
N/A
### Operating system version
na
### Additional context
_No response_
### Issue checklist
- [X] I searched for previous requests in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=)
- [X] I verified the feature was still missing when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds)
- [X] My request contains all necessary details
|
non_process
|
boost libraries with the exact match describe the request when i search for sd i expect to see to show up in the top of the list so i can install it without a lot of scrolling through the list of results but instead it is far down in the results related to describe the current behavior arduino cli lib search sd as you can see the sd lib is pretty down in the list arduino cli version operating system n a operating system version na additional context no response issue checklist i searched for previous requests in i verified the feature was still missing when using the my request contains all necessary details
| 0
|
35,884
| 2,793,805,886
|
IssuesEvent
|
2015-05-11 13:32:23
|
SarineTechnologies/sarine.viewer.templates
|
https://api.github.com/repos/SarineTechnologies/sarine.viewer.templates
|
opened
|
Design issues
|
High priority Template v.1.10.0
|
issues to resolve:
https://projects.invisionapp.com/share/MG276SMQW#/screens
1. "Hearts & Arrows" at the top should be broken to 2 lines.
2. Distance between (at the top) H&A and cut should be shortened.
3. Question mark should be blue (it may be blue, i am unable to determine if it is blue or purple).
4. In the cut screen, text should be "View the accurate geometrical measurements of the diamond to evaluate the proportions." (not like in the design).
|
1.0
|
Design issues - issues to resolve:
https://projects.invisionapp.com/share/MG276SMQW#/screens
1. "Hearts & Arrows" at the top should be broken to 2 lines.
2. Distance between (at the top) H&A and cut should be shortened.
3. Question mark should be blue (it may be blue, i am unable to determine if it is blue or purple).
4. In the cut screen, text should be "View the accurate geometrical measurements of the diamond to evaluate the proportions." (not like in the design).
|
non_process
|
design issues issues to resolve hearts arrows at the top should be broken to lines distance between at the top h a and cut should be shortened question mark should be blue it may be blue i am unable to determine if it is blue or purple in the cut screen text should be view the accurate geometrical measurements of the diamond to evaluate the proportions not like in the design
| 0
|
9,776
| 12,794,135,735
|
IssuesEvent
|
2020-07-02 06:10:59
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Linux and Windows Service Containers
|
Pri2 devops-cicd-process/tech devops/prod product-question
|
Our team wants to run steps within a Linux container that uses a Windows service container (https://hub.docker.com/r/microsoft/azure-cosmosdb-emulator)
Do you have any advice for how we can achieve this today? Is it something that will be possible in the future?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7d1c794c-d676-fc7e-758e-3464f16a1554
* Version Independent ID: 941de04a-2d34-ec82-8612-00e4070de673
* Content: [Service Containers - Azure Pipelines & TFS](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/service-containers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/service-containers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/service-containers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Linux and Windows Service Containers - Our team wants to run steps within a Linux container that uses a Windows service container (https://hub.docker.com/r/microsoft/azure-cosmosdb-emulator)
Do you have any advice for how we can achieve this today? Is it something that will be possible in the future?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7d1c794c-d676-fc7e-758e-3464f16a1554
* Version Independent ID: 941de04a-2d34-ec82-8612-00e4070de673
* Content: [Service Containers - Azure Pipelines & TFS](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/service-containers?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/service-containers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/service-containers.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
linux and windows service containers our team wants to run steps within a linux container that uses a windows service container do you have any advice for how we can achieve this today is it something that will be possible in the future document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
5,392
| 8,214,905,781
|
IssuesEvent
|
2018-09-05 02:11:09
|
googleapis/nodejs-translate
|
https://api.github.com/repos/googleapis/nodejs-translate
|
closed
|
There no named export "Translate"
|
type: process
|
Here => https://github.com/googleapis/nodejs-translate/blob/master/samples/quickstart.js#L20
The quickstart is wrong, it should be:
```
const Translate = require('@google-cloud/translate');
```
|
1.0
|
There no named export "Translate" - Here => https://github.com/googleapis/nodejs-translate/blob/master/samples/quickstart.js#L20
The quickstart is wrong, it should be:
```
const Translate = require('@google-cloud/translate');
```
|
process
|
there no named export translate here the quickstart is wrong it should be const translate require google cloud translate
| 1
|
486,308
| 14,007,165,724
|
IssuesEvent
|
2020-10-28 21:07:03
|
AOEpeople/crawler
|
https://api.github.com/repos/AOEpeople/crawler
|
opened
|
[BUG] Crawler return invalid content/cache to StaticFileCache
|
3rd party ext Bug Priority 1 TYPO3v10 TYPO3v9 refactoring
|
## Bug Report
This bug is still WIP in regard to describing it. But to not lose the problem I'll write down what I have for now.
There is an issue on the Static File Cache GitHub on Crawler Compatibility:
https://github.com/lochmueller/staticfilecache/issues/260
The crawler appears to have some issues with the Middleware Handling and the content/caching that is handled over to the StaticFileCache.
This is currently resulting in an Invalid cache in Static File Cache. This is since commit: https://github.com/lochmueller/staticfilecache/commit/975eff6b95c712e34df6ff8faad8be20ef51032a omitted by a warning in regard to the crawler.
As we don't want to break functionality of other extensions, and of course not lose users of the Crawler we will try to have this fixed.
If you have any information that could be helpful to solve this issue, please add a comment below and lets see how we can best solve this issue.
|
1.0
|
[BUG] Crawler return invalid content/cache to StaticFileCache - ## Bug Report
This bug is still WIP in regard to describing it. But to not lose the problem I'll write down what I have for now.
There is an issue on the Static File Cache GitHub on Crawler Compatibility:
https://github.com/lochmueller/staticfilecache/issues/260
The crawler appears to have some issues with the Middleware Handling and the content/caching that is handled over to the StaticFileCache.
This is currently resulting in an Invalid cache in Static File Cache. This is since commit: https://github.com/lochmueller/staticfilecache/commit/975eff6b95c712e34df6ff8faad8be20ef51032a omitted by a warning in regard to the crawler.
As we don't want to break functionality of other extensions, and of course not lose users of the Crawler we will try to have this fixed.
If you have any information that could be helpful to solve this issue, please add a comment below and lets see how we can best solve this issue.
|
non_process
|
crawler return invalid content cache to staticfilecache bug report this bug is still wip in regard to describing it but to not lose the problem i ll write down what i have for now there is an issue on the static file cache github on crawler compatibility the crawler appears to have some issues with the middleware handling and the content caching that is handled over to the staticfilecache this is currently resulting in an invalid cache in static file cache this is since commit omitted by a warning in regard to the crawler as we don t want to break functionality of other extensions and of course not lose users of the crawler we will try to have this fixed if you have any information that could be helpful to solve this issue please add a comment below and lets see how we can best solve this issue
| 0
|
2,200
| 5,040,616,618
|
IssuesEvent
|
2016-12-19 06:41:53
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
opened
|
[subtitles] [eng] A bas les paradis fiscaux ! J L Mélenchon au Parlement européen
|
Language: English Process: [1] Writing in progress
|
# Video title
A bas les paradis fiscaux ! J L Mélenchon au Parlement européen
# URL
https://www.youtube.com/watch?v=2TbjWeOxQNg&t=4s
Youtube subtitle language
Anglais
Duration
1:08
URL subtitles
https://www.youtube.com/timedtext_editor?lang=en&v=2TbjWeOxQNg&tab=captions&ref=player&action_mde_edit_form=1&ui=hd&bl=vmp
|
1.0
|
[subtitles] [eng] A bas les paradis fiscaux ! J L Mélenchon au Parlement européen - # Video title
A bas les paradis fiscaux ! J L Mélenchon au Parlement européen
# URL
https://www.youtube.com/watch?v=2TbjWeOxQNg&t=4s
Youtube subtitle language
Anglais
Duration
1:08
URL subtitles
https://www.youtube.com/timedtext_editor?lang=en&v=2TbjWeOxQNg&tab=captions&ref=player&action_mde_edit_form=1&ui=hd&bl=vmp
|
process
|
a bas les paradis fiscaux j l mélenchon au parlement européen video title a bas les paradis fiscaux j l mélenchon au parlement européen url youtube subtitle language anglais duration url subtitles
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.