added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T06:38:14.742080
| 2022-10-15T16:17:00
|
1410216622
|
{
"authors": [
"harshraj8843",
"roberanegussie"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4818",
"repo": "codinasion/program",
"url": "https://github.com/codinasion/program/issues/4405"
}
|
gharchive/issue
|
Write a C# program to convert a string to an integer
Description
Write a C# program to convert a string to an integer
Input : "123"
Output : 123
How to contribute
Save the solution in program/ConvertAStringToAnInteger.cs file
Add ConvertAStringToAnInteger.cs file in convert-a-string-to-an-integer folder
!assign
Hey @roberanegussie, this issue is already assigned to @anandfresh !!!
Please choose another issue.
Thanks for your interest in contributing to this project.
|
2025-04-01T06:38:14.746210
| 2017-05-24T16:02:42
|
231091215
|
{
"authors": [
"abhishek97",
"akansh97531",
"nikhilkumarsingh",
"sourabhtk37"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4819",
"repo": "coding-blocks/content-downloader",
"url": "https://github.com/coding-blocks/content-downloader/issues/2"
}
|
gharchive/issue
|
Create desktop GUI
ctdl currently works as a command line utility.
Create a desktop GUI for ctdl using Tkinter.
Is using Tkinter compulsory? Can't we use other tools for this? Like electron?https://electron.atom.io
Well, I have no idea if it is easy/possible to integrate python scripts with electron.
Tkinter is not compulsory though. You can try electron if it is possible.
It is possible to use electron here. The UI/UX with Tkinter is a bust.
Heres how it'll go using electron:
Write a Python Wrapper to expose the cli as an API for the electron app to consume. (something like zeromq )
Use electron to server static files (.html) and use the api.
In the process we will have a web api for the same as well.
Ok then. Go for it!
Wouldn't that be too complex @abhishek97 ? (import this)
(totally keeping aside electron's aspects)
Now, if you are talking about cross platform aspect then why not kivy?
Why the need of desktop GUI than a simple webapp? @nikhilkumarsingh
#5 gui for content downloader @nikhilkumarsingh
|
2025-04-01T06:38:14.759102
| 2021-02-27T06:36:20
|
817850292
|
{
"authors": [
"coding-horror",
"drewjcooper"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4820",
"repo": "coding-horror/basic-computer-games",
"url": "https://github.com/coding-horror/basic-computer-games/pull/97"
}
|
gharchive/pull-request
|
84 Super Star Trek CSharp
WIP
Super Trek Trek port to C#
Introduction including instructions
Main game loop
Short Range Scan
Shield Control
More to come...
This is a big one, take your time! Thanks!
|
2025-04-01T06:38:14.768987
| 2023-09-19T17:10:54
|
1903409099
|
{
"authors": [
"gdamou",
"neolectron"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4821",
"repo": "codinglab-io/discord-bot",
"url": "https://github.com/codinglab-io/discord-bot/issues/92"
}
|
gharchive/issue
|
Can't display commands due to insufficient permissions
When writing "/" we don't have all the commands displaying :
This is clearly a non intentionall bug.
/fart should be available for everyone to use.
I think the scope of this fix can be greater tho. I believe we should have a way to declare permissions for our slash commands, which we do not have yet.
Interesting topic that need to be discussed.
Thanks for your report!
|
2025-04-01T06:38:14.772858
| 2017-09-05T05:43:10
|
255165332
|
{
"authors": [
"mutedbytes",
"rivaldi8"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4822",
"repo": "codinguser/gnucash-android",
"url": "https://github.com/codinguser/gnucash-android/issues/726"
}
|
gharchive/issue
|
Asset accounts gets (-) negative sign prepended when editing, causing double negative
Steps to reproduce the behaviour
Create a transaction to decrease amount in an asset account.
Click to edit the transaction, a (-) sign appears in front of red decreasing value
Click to save the edited transaction
-> error = the "decrease" transaction now becomes a "negative decrease" resulting in a positive increase to the asset balance
This behavior occurred after changing default transactions to be "credit" rather than "debit" in settings. Not sure if this is relevant.
Expected behaviour
Expected behavior would be to have the negative not included in the calculation.
Actual behaviour
Negative sign seems to be prepended upon a decrease transaction type. After editing and saving, the negative seems to be included in the resulting calculations, causing the transaction to increase the asset balance rather than decreasing it as should be.
Software specifications
GnuCash Android version: 2.2.1
System Android version: 6.0.1
Device type: LG Nexus 5
This has already been reported in #723. Please, add any further comments there. Thanks!
|
2025-04-01T06:38:14.777049
| 2020-10-30T11:48:15
|
733109403
|
{
"authors": [
"boaz-codota",
"elovin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4823",
"repo": "codota/tabnine-vscode",
"url": "https://github.com/codota/tabnine-vscode/issues/179"
}
|
gharchive/issue
|
De-prioritized TabNine suggestions to be at the end of the list
please complete the following information:
OS version: ArchLinux kernel 5.9.1
Editor version: 1.50.1
Programming language: php,TypeScript,bash
TabNine extension version: 2.8.8
Issue Details:
This is a feature request
The current official TabNine is unusable for my colleagues and myself because we can not change the suggestion priority.
We need the priority to come after the build in suggestions otherwise we get less useful suggestions at the top most of the time.
This problem is faced by a lot of people since there is an unofficial TabNine extension for vscode which does just that.
With that feature TabNine is a great tool to reduce repetitive typing tasks.
So can you please take look at this unofficial extension and add a way to change the suggestion priority ?
Hey @elovin thanks for reaching out to us!
Thanks for the suggestion, we will try to address that in upcoming releases.
Thanks, Boaz.
|
2025-04-01T06:38:14.794545
| 2019-05-28T23:06:34
|
449515691
|
{
"authors": [
"Emuentes",
"GeorgeWL",
"codypearce"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4824",
"repo": "codypearce/material-bread",
"url": "https://github.com/codypearce/material-bread/issues/186"
}
|
gharchive/issue
|
Request: Implement a Searchable Select
I think it would be nice to have a Select with Input component, I've seen it used in many Google Material products for making it easier to filter large lists.
Something like this little library: https://selectize.github.io/selectize.js/
I'm by no means expecting you to dictate the behaviour of the select, in fact I'd rather you didn't so that one could do so themselves, or at least add an optional onChange prop override.
Yeah under the hood the Select component using the TextField component https://github.com/codypearce/material-bread/blob/master/src/Components/Select/Select.js#L87, but it needs some rewriting to be able to type inside it and activate the dropdown.
I think there's two parts to that:
A prop to distinguish between allowing the user to type in the textfield and only allowing them to click to activate the dropdown.
Better support for chips in TextFields, like here https://material.io/design/components/chips.html#
What do you think?
I agree with that yeah.
Possibly have the prop be boolean and something like filterable ?
Another possibility is to allow it to have children, but that could lead to behaviours against the material ethos.
Well I like the idea of of prebuilt functionality that matches material, but with enough escape hatches to make something more custom.
So maybe add that filterable prop, but allow for the user to pass in a custom TextField component that gets rendered instead of the prebuilt one, for example renderTextField prop. This would also allow the user to use other packages like https://github.com/benhurott/react-native-masked-text while still taking advantage of the Select element.
Yeah that sounds great
What about downshift?
Might not need too much refactoring to be cross platform.
I'm not sure tho, haven't looked into it.
https://github.com/downshift-js/downshift
|
2025-04-01T06:38:14.799290
| 2017-09-11T16:34:22
|
256766483
|
{
"authors": [
"dchenier",
"j-rewerts"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4825",
"repo": "coe-google-apps-support/Mailman",
"url": "https://github.com/coe-google-apps-support/Mailman/issues/168"
}
|
gharchive/issue
|
Shaylen Naidoo: Conditional Send Should Support Logical Operations
Email
<EMAIL_ADDRESS>Description
Allow sending of mail merges to use several logical conditions such that we can use operations like AND, OR and NOT to join the multiple logical conditions.
This is now possible in Rich Text Mailman, because all of the fields are rendered using the Handlebars templating engine. (to, cc, bcc, subject, body, and condition)
For example, If you have two columns "My Condition 1" and "My Condition 2":
To only send emails when both are TRUE use {{#and [My Condition 1] [My Condition 2]}}
To send emails if either are TRUE use {{#or [My Condition 1] [My Condition 2]}}
More powerful logical operators exist, documentation will be forthcoming
|
2025-04-01T06:38:14.807985
| 2021-09-02T06:03:52
|
986196915
|
{
"authors": [
"MrOrz",
"coveralls"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4826",
"repo": "cofacts/rumors-line-bot",
"url": "https://github.com/cofacts/rumors-line-bot/pull/286"
}
|
gharchive/pull-request
|
Set cache-control for assets
As discussed in slack, after we remove max-age in #284, cloudflare never caches our asset files.
This is because
koa-send (used by koa-static-server we are currently using) sets max-age to 0 if we do not specify any max-age: https://github.com/koajs/send/blob/master/index.js#L60
Default cache behavior of cloudflare will respect max-age: 0 when it is given
In this PR we override Cache-Control using setHeaders option so that whenever koa-send serves a file that is not index.html, we attach a 1-year long max-age in the response header.
HTML
Assets that is not HTML
Pull Request Test Coverage Report for Build<PHONE_NUMBER>
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 87.326%
Totals
Change from base Build<PHONE_NUMBER>:
0.0%
Covered Lines:
968
Relevant Lines:
1094
💛 - Coveralls
|
2025-04-01T06:38:14.813301
| 2024-03-31T16:45:26
|
2217078408
|
{
"authors": [
"coffeebeats"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4827",
"repo": "coffeebeats/gdbuild",
"url": "https://github.com/coffeebeats/gdbuild/pull/81"
}
|
gharchive/pull-request
|
chore: release v0.2.4
:robot: I have created a release beep boop
0.2.4 (2024-04-01)
What's Changed
feat(ci): pre-build arm64 on linux binaries by @coffeebeats in https://github.com/coffeebeats/gdbuild/pull/80
fix(scripts): unblock downloads of new arm64 on linux target by @coffeebeats in https://github.com/coffeebeats/gdbuild/pull/82
fix(scripts): use correct compound condition syntax by @coffeebeats in https://github.com/coffeebeats/gdbuild/pull/83
feat(ci): add support for explicit --debug flag by @coffeebeats in https://github.com/coffeebeats/gdbuild/pull/84
Full Changelog: https://github.com/coffeebeats/gdbuild/compare/v0.2.3...v0.2.4
This PR was generated with Release Please. See documentation.
:robot: Release is at https://github.com/coffeebeats/gdbuild/releases/tag/v0.2.4 :sunflower:
|
2025-04-01T06:38:14.821137
| 2024-03-11T09:04:40
|
2178603351
|
{
"authors": [
"crusaderky",
"milesgranger"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4828",
"repo": "coiled/benchmarks",
"url": "https://github.com/coiled/benchmarks/pull/1447"
}
|
gharchive/pull-request
|
Migrate AB and CI dependencies off conda to pip-compile
Will close #1437
#1437
Moves dependencies to new AB_<name>.requirements.txt file, which is dependabot friendly for automated upgrades. Alone w/ the associated dependabot.yml file configured to look in AB_environments directory.
Successful A/B run here: https://github.com/coiled/benchmarks/actions/runs/8246636881
@crusaderky, it seemed to get unwieldly needing to copy around a core set of requirements to so many files when moving to pip. Therefore I've converted to pip-compile so we can include these in other environments, by just adding -r requirements.in or similar both in CI and AB environments. Then just two conda env files, one for the base to determine python version and stuff like openssl and openjdk which is easier w/ conda, then one for updating to git-tip.
Done in https://github.com/coiled/benchmarks/pull/1447/commits/d628ccee9253acb9f3bff8f332bf2643761a6c40, let me know what you think.
@milesgranger I've gone through all the documentation; please review.
If you're happy with it I think we can merge?
|
2025-04-01T06:38:14.991372
| 2022-01-13T15:37:26
|
1101948515
|
{
"authors": [
"Svisstack",
"jamieyello"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4829",
"repo": "coinapi/coinapi-sdk",
"url": "https://github.com/coinapi/coinapi-sdk/issues/125"
}
|
gharchive/issue
|
C# option to toggle debug output
Writing to the console multiple times a frame is actually a pretty demanding task, and it clutters up other outputs I'm looking for while debugging. Seeing the output of every string that comes in, while cool, keeps me from seeing more important stuff, and while it's useful for knowing what kind of information you're receiving there should be a way to toggle it on or off.
Fixed in #126. You can basically Debug on the handler side of the event if you want. Upgrade nuget to 2.0.1
|
2025-04-01T06:38:14.997294
| 2020-09-12T00:39:48
|
699904087
|
{
"authors": [
"cindyxkuang",
"coveralls"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4830",
"repo": "coinbase/rosetta-cli",
"url": "https://github.com/coinbase/rosetta-cli/pull/130"
}
|
gharchive/pull-request
|
improved error handling and test coverage
Fixes # .
Motivation
we've updated the Rosetta SDK to include named error types for better visibility on error handling. we're updating the CLI here to leverage that new functionality to improve coverage in block syncing and balance checking tests.
Solution
we are using the Err() functions that the SDK now exposes to check if an error originated from the storage or syncer packages.
we also make sure to fail noisily by having check:data hard exit if we encounter an error not handled by one of our tests.
Open questions
Pull Request Test Coverage Report for Build 4571
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 84.737%
Totals
Change from base Build 4550:
0.0%
Covered Lines:
161
Relevant Lines:
190
💛 - Coveralls
|
2025-04-01T06:38:15.015716
| 2021-11-11T07:54:45
|
1050675380
|
{
"authors": [
"antstorm",
"duy-chk"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4831",
"repo": "coinbase/temporal-ruby",
"url": "https://github.com/coinbase/temporal-ruby/issues/116"
}
|
gharchive/issue
|
Trying to run an example but failed with DNS resolution failed for service error
I tried to run temporal locally and try with hello world workflow but get this error. Could anyone help me?
Temporal.start_workflow(HelloWorldWorkflow)
[DEPRECATION] This method is now deprecated without a substitution
GRPC::Unavailable: 14:DNS resolution failed for service: :. debug_error_string:{"created":"@1636617011.591632000","description":"Resolver transient failure","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":1356,"referenced_errors":[{"created":"@1636617011.591631000","description":"DNS resolution failed for service: :","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/dns_resolver_ares.cc","file_line":360,"grpc_status":14,"referenced_errors":[{"created":"@1636617011.591625000","description":"unparseable host:port","file":"src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc","file_line":843,"target_address":":"}]}]}
from /Users/duy-chk/.rbenv/versions/2.7.2/lib/ruby/gems/2.7.0/gems/grpc-1.41.1-universal-darwin/src/ruby/lib/grpc/generic/active_call.rb:29:in `check_status'
I am able to run temporal server and web service. Also run the configuration block already.
Temporal.configure do |config|
config.host = 'localhost'
config.port = 7233
config.namespace = 'ruby-samples'
config.task_queue = 'hello-world'
end
Since you've closed the issue I'm assuming this has been resolved, please let me know if that's not the case and you need help
Hi @antstorm, thanks for your reply. It was resolved by restarting irb.
|
2025-04-01T06:38:15.020374
| 2020-05-26T10:34:37
|
624780705
|
{
"authors": [
"tim2CF"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4833",
"repo": "coingaming/lnd-client",
"url": "https://github.com/coingaming/lnd-client/issues/29"
}
|
gharchive/issue
|
RawConfig FromJSON
Sometimes when we are operating with multiple LND nodes in real apps, it's convenient to read LND environment from some JSON environment variable
refactor RawConfig fields types (proper types instead of blind ByteString and Int)
implement FromJSON type class for RawConfig type and its fields
use smart constructors for FromJSON instances of fields where it is possible
if value is just newtype without smart constructor, use GeneralizedNewtypeDeriving
Let's also rename type to RawLndEnv and export it (just type, not constructors because it supposed to be read from JSON or environment variables, not constructed directly)
done 🚀
|
2025-04-01T06:38:15.021780
| 2020-08-31T15:18:05
|
689287750
|
{
"authors": [
"colbyfayock",
"zacjones93"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4834",
"repo": "colbyfayock/launchtime-workshop",
"url": "https://github.com/colbyfayock/launchtime-workshop/pull/3"
}
|
gharchive/pull-request
|
Add Exercise 03 Extra Credit - Env Vars
This exercise listed a TODO so I took a stab at adding it based on the lesson you recorded.
thanks Zac!!
|
2025-04-01T06:38:15.029750
| 2019-10-23T23:53:38
|
511629436
|
{
"authors": [
"codecov-io",
"dirk-thomas",
"rotu"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4835",
"repo": "colcon/colcon-core",
"url": "https://github.com/colcon/colcon-core/pull/258"
}
|
gharchive/pull-request
|
More Pythonic EventReactor
Make EventReactor a context manager
Use Thread via composition instead of inheritance
Codecov Report
Merging #258 into master will increase coverage by 0.03%.
The diff coverage is 88.88%.
@@ Coverage Diff @@
## master #258 +/- ##
==========================================
+ Coverage 80.04% 80.07% +0.03%
==========================================
Files 54 54
Lines 3122 3127 +5
Branches 518 518
==========================================
+ Hits 2499 2504 +5
Misses 584 584
Partials 39 39
Impacted Files
Coverage Δ
colcon_core/event_reactor.py
100% <100%> (ø)
:arrow_up:
colcon_core/executor/__init__.py
95.23% <81.81%> (-0.07%)
:arrow_down:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update c5983df...819b170. Read the comment docs.
Thanks for the contribution.
You’re always welcome!
|
2025-04-01T06:38:15.054248
| 2024-08-18T11:47:29
|
2471929185
|
{
"authors": [
"sheepbox8646"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4836",
"repo": "colinhacks/zod",
"url": "https://github.com/colinhacks/zod/issues/3715"
}
|
gharchive/issue
|
z.ref
There is a special usage in JSON Schema:
{
"$schema": "http://json-schema.org/draft-07/schema#",
"properties": {
"root": {
"type": "object",
"properties": {
"type": {
"type": "string"
},
"arguments": {
"type": "array"
},
"children": {
"type": "array",
"items": {
"$ref": "#/properties/root"
}
},
}
}
},
"required": [
"root"
]
}
$ref means use a defined structure in the tree.
Have zod a similar function or not? If not, there is any posibility add it?
I try to use a defined structure in Zod, but the ts-server says the types are loop used
|
2025-04-01T06:38:15.068508
| 2018-11-28T17:13:48
|
385372221
|
{
"authors": [
"olivx",
"thijstriemstra"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4837",
"repo": "collab-project/videojs-record",
"url": "https://github.com/collab-project/videojs-record/issues/308"
}
|
gharchive/issue
|
error to play recorded data on firefox
Description
I get erro on firefox , when i finish record data and go to playback in de same video tag.
on chrome it's work.
i don't understand the erro, you could help me ?
Steps to reproduce
const timeout = 30
var options = {
controls: false,
width: 554,
height: 240,
fluid: false,
controlBar: {
fullscreenToggle: false,
volumePanel: false
},
plugins: {
record: {
videoMimeType: 'video/webm',
audio: true,
video: {
mandatory: {
minWidth: 320,
minHeight: 240,
},
},
maxLength: timeout,
debug: true
}
}
};
var player = videojs('record_retc', options, function() {
// print version information at startup
var msg = 'Using video.js ' + videojs.VERSION +
' with videojs-record ' + videojs.getPluginVersion('record') +
' and recordrtc ' + RecordRTC.version;
videojs.log(msg);
});
const sleep = (milliseconds) => {
return new Promise(resolve => setTimeout(resolve, milliseconds))
}
document.querySelector('.start_record').onclick = function(event){
player.record().getDevice();
sleep(1 * 1000).then(() => {
player.record().start()
})
}
player.on('stopRecord', function(event){
console.log('stopped recording!');
player.record().stopDevice()
})
player.on('finishRecord', function() {
console.log('finished recording:', player.recordedData);
console.log(player.record().getDuration())
player.on('loadeddata' , function() {
console.log('starting playback');
player.play();
});
var blobUrl = window.URL.createObjectURL(player.recordedData.video);
player.src({type: player.recordedData.video.type, src: blobUrl});
});
console log
Using recorderType: MediaStreamRecorder RecordRTC.js:1039:9
Passing following config over MediaRecorder API.
Object { type: "video", video: {…}, canvas: {…}, frameInterval: 10, disableLogs: false, recorderType: null, mimeType: "video/webm", timeSlice: undefined, onTimeStamp: undefined, initCallback: initCallback()
, … }
RecordRTC.js:2043:13
Recorder state changed: recording RecordRTC.js:699:17
Initialized recorderType: MediaStreamRecorder for output-type: video RecordRTC.js:99:13
started recording! test:432:5
stopped recording! test:436:3
Stopped recording video stream. RecordRTC.js:125:13
Recorder state changed: stopped RecordRTC.js:699:17
video/webm -> 3.90 MB RecordRTC.js:166:17
finished recording:
Blob { lastModified:<PHONE_NUMBER>691, lastModifiedDate: Date 2018-11-28T16:58:08.691Z, name: "1543424288691.webm", size: 3904337, type: "video/webm" }
Results
Expected
Please describe what you expected to see.
Actual
Please describe what actually happened.
Error output
VIDEOJS: ERROR: TypeError: "right-hand side of 'in' should be an object, got undefined"
createObjectURL https://webrtc.github.io/adapter/adapter-latest.js:4014:7
http://<IP_ADDRESS>:8000/pt-br/jobconvo/NTc1Ng-analista-trade-marketing/23778629-2214-410c-889f-cd12bb2d8c96/test/:456:19
bound https://vjs.zencdn.net/7.3.0/video.js:2168:14
dispatcher https://vjs.zencdn.net/7.3.0/video.js:1818:17
trigger https://vjs.zencdn.net/7.3.0/video.js:1954:7
1 https://vjs.zencdn.net/7.3.0/video.js:2832:14
value https://cdnjs.cloudflare.com/ajax/libs/videojs-record/2.4.1/videojs.record.min.js:8:17415
bound https://vjs.zencdn.net/7.3.0/video.js:2168:14
dispatcher https://vjs.zencdn.net/7.3.0/video.js:1818:17
trigger https://vjs.zencdn.net/7.3.0/video.js:1954:7
1 https://vjs.zencdn.net/7.3.0/video.js:2832:14
value https://cdnjs.cloudflare.com/ajax/libs/videojs-record/2.4.1/videojs.record.min.js:8:41484
getBlob https://webrtcexperiment-webrtc.netdna-ssl.com/RecordRTC.js:1348:13
value https://cdnjs.cloudflare.com/ajax/libs/videojs-record/2.4.1/videojs.record.min.js:8:41104
stopRecording https://webrtcexperiment-webrtc.netdna-ssl.com/RecordRTC.js:1264:17
_callback https://webrtcexperiment-webrtc.netdna-ssl.com/RecordRTC.js:173:21
ondataavailable https://webrtcexperiment-webrtc.netdna-ssl.com/RecordRTC.js:2123:17
video.js:142:49
Additional Information
Please include any additional information necessary here. Including the following:
versions
videojs
what version of videojs does this occur with?
browsers
what browser(s) are affected? Make sure to test with all third-party browser extensions disabled.
OSes
what platforms (operating systems and devices) are affected?
You do not need to load the data manually, it loads automatically after you stop recording, see examples in repository.
sorry,
but loads automatically after i stop recording, don't work , what's exemplo do that ?
i need config some thing ?
ps: i don't using the controls , i stop and play manually
I see. What happens when you use this in the finishRecord handler (player.src is not supported afaik):
player.record().load(blobUrl);
@olivx seems you're also running into the issue where you have to use player.recordedData.video on Chrome and player.recordedData on Firefox. This will be fixed in v3.0.0 (see #270).
worked this way....
if (navigator.userAgent.toLowerCase().indexOf("chrome") != -1){
console.log('google chrome')
blobUrl = window.URL.createObjectURL(player.recordedData.video);
player.src({type: player.recordedData.video.type, src: blobUrl});
}
if (navigator.userAgent.toLowerCase().indexOf("firefox") != -1){
console.log('mozila firefox')
blobUrl = window.URL.createObjectURL(player.recordedData);
player.src({type: player.recordedData.type, src: blobUrl});
}
thanks a lot for your help !
No worries! In 3.0.0 it should be simply blobUrl = window.URL.createObjectURL(player.recordedData);.
|
2025-04-01T06:38:15.131977
| 2023-12-12T17:08:17
|
2038210524
|
{
"authors": [
"cydneyj303",
"falquaddoomi",
"vincerubinetti"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4838",
"repo": "colorado-cancer-center/COCancerScope",
"url": "https://github.com/colorado-cancer-center/COCancerScope/issues/20"
}
|
gharchive/issue
|
Say where sources came from
Split from #15
Say where sources came from. Defer to CancerInFocus's sources page rather than duplicating data.
We're going to have a dedicated Sources page, accessible from the header of the site, and populated with static text and citation data about each source, including timespans (per #21 ), linking to external info pages when possible/appropriate.
A shortened label or id or some citation for the source will be hardcoded into the backend for now and displayed on the frontend in the legend.
From Sean:
Perhaps start as a google doc to keep list of where things come from and provide user-level and technical documentation.
From the meeting:
@vincerubinetti : would be good to have a meeting where we interactively put this together.
Merging in #21
Also specify timespans of data.
I believe the CancerInFocus data is from the last 2-5 years or something like that? Might be best to defer external pages for that info? Maybe linking to the appropriate documentation that says the timespans is enough...?
Hi all -
Attached is a document with our data sources page. There is one highlighted part that we need your team's eyes for. If the highlighted part remains true then we can keep it in and can publish to the platform. If it is not true, we can delete or edit and publish.
Thanks!
DATA SOURCES_Final.docx
|
2025-04-01T06:38:15.136602
| 2022-06-07T18:58:08
|
1263734696
|
{
"authors": [
"coloradocolby",
"piotr-machura"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4839",
"repo": "coloradocolby/fsrx",
"url": "https://github.com/coloradocolby/fsrx/issues/1"
}
|
gharchive/issue
|
Newlines stripped when reading STDIN
Description
When the file is provided as an argument, the newlines are handled correctly. But when fsrx reads from standard input the newlines are stripped away.
To Reproduce
$ fsrx .bashrc # as expected
$ cat .bashrc | fsrx # a mess
Expected behavior
The newline handling should be consistent between those two methods.
Screenshots
Additional context
Fixing this would make it possible to use fmt to reflow incoming text to a managable width, like so:
$ fmt -w 60 long_lines.txt | fsrx | less -R
Right now the newlines created by fmt are deleted.
Just noticed that (some?) special characters at the end of lines are stripped as well. - notice the dissappearing ! in the screenshot above.
Here is another screenshot with a line from running fsrx ~/.bashrc, notice the unclosed parenthesis and double quotes:
Is this by design?
@piotr-machura hmm I know this was working at some point, I must have added a regression. Looking now!
all fixed! releasing now
|
2025-04-01T06:38:15.169899
| 2024-11-05T22:13:53
|
2636578511
|
{
"authors": [
"0xnm",
"himanshumahajan138",
"lihaoyi"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4840",
"repo": "com-lihaoyi/mill",
"url": "https://github.com/com-lihaoyi/mill/pull/3913"
}
|
gharchive/pull-request
|
Improve Android SDK handling
This PR seeks to improve Android SDK handling approach.
The current issues are:
Android SDK setup is limited only to Linux and is also using tooling which may be not available on other platforms.
Each module will have its own copy of SDK. Imagine a project with 20 Android modules (19 lib modules + 1 app module), it will be 20x space usage than it should be. This is quite significant, considering that the basic setup (platforms:android-35, platform-tools and build-tools;35.0.0) is taking ~300 MB.
Licenses are accepted on the end-user behalf, without the consent.
This PR:
Aligns the SDK setup experience with AGP (Android Gradle Plugin):
Expects Android SDK (at least command line tools) to be present on the end-user machine and registered with the env variable (ANDROID_HOME).
Expects command line tools to be present there to utilize sdkmanager to setup the necessary SDK components.
Expects user to accept the necessary licenses by themselves. If they are not accepted, sdkmanager will block stdin waiting for the user input before proceeding with installation. Note: when downloading Android Studio, it will come with the necessary basic license already accepted + some basic SDK components. If downloading command line tools, it is necessary to accept the main license (android-sdk-license) manually.
Installs all the necessary components to the location pointed by the ANDROID_HOME env variable, so that they are shared by all modules.
Adds the necessary step on the CI to install Android SDK for the jobs where it is needed. Note: I was also thinking about extracting Android-related examples into a dedicated job, but then for the existing jobs doing something like example.kotlinlib[__].local.testCached ->example.kotlinlib[__:^AndroidAppKotlinModule].local.testCached doesn't work (at least mill resolve still shows Android module), all cross-segments should be listed manually.
Speaking of sdkmanager: it is also possible to utilize directly the following artifact and put it in a worker, but for this one it is impossible to find a version on the Android Developers website, so end-user will have troubles updating it and anyway it is one-shot action normally, so keeping it even in the separate classpath looks like an overkill. Command line tools will be needed anyway further for the other tools it provides.
By some reason CI fails (while there is an installation of SDK components in the logs), I will investigate.
@0xnm Sir your approach is very good but i want to add something that i did during kotlin example please once see this pr #3769
here please see the AndroidAppModule and AndroidSdkModule
main points:
used cache for downloaded setup
updated the code for android setup
actually in that PR i have done a alot of progress but the only thing i am stucked with library resources (aar files packaging) and almost all things are done i have updated the setup for cache and also updated the android app creation process but i was unable to solve the addition of jetpack compose fully(only aar packaging left)
i request you to please once have a look and take some insights i think this will help...
That was my miss, I forgot to add the line https://github.com/com-lihaoyi/mill/pull/3913/files#diff-7314d0ebbd2e9537ae4889316745b4fd2fa43cb86275c9caae18a86ba228b642R96, CI is fine now.
Thanks! Will take a look
|
2025-04-01T06:38:15.187094
| 2024-02-13T00:35:14
|
2131235034
|
{
"authors": [
"ashishchandr70",
"hvanz",
"melekes"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4841",
"repo": "cometbft/cometbft",
"url": "https://github.com/cometbft/cometbft/issues/2319"
}
|
gharchive/issue
|
State Sync not working
I am testing state-sync from our RPC nodes which have been set up according to the config described below. Please let me know if I am doing anything wrong, or leaving out any necessary step.
Here is my application's tendermint version:
ABCI: 1.0.0
BlockProtocol: 11
P2PProtocol: 8
Tendermint: 0.37.1
My RPC nodes have the following config in app.toml
###############################################################################
### State Sync Configuration ###
###############################################################################
# State sync snapshots allow other nodes to rapidly join the network without replaying historical
# blocks, instead downloading and applying a snapshot of the application state at a given height.
[state-sync]
# snapshot-interval specifies the block interval at which local state sync snapshots are
# taken (0 to disable).
snapshot-interval = 1000
# snapshot-keep-recent specifies the number of recent snapshots to keep and serve (0 to keep all).
snapshot-keep-recent = 2
The node I am trying to bring up via state-sync has the following in config.toml. Trust Height is latest block -1000 and is updated each time, along with trust hash.
#######################################################
### State Sync Configuration Options ###
#######################################################
[statesync]
# State sync rapidly bootstraps a new node by discovering, fetching, and restoring a state machine
# snapshot from peers instead of fetching and replaying historical blocks. Requires some peers in
# the network to take and serve state machine snapshots. State sync is not attempted if the node
# has any local state (LastBlockHeight > 0). The node will have a truncated block history,
# starting from the height of the snapshot.
enable = true
# RPC servers (comma-separated) for light client verification of the synced state machine and
# retrieval of state data for node bootstrapping. Also needs a trusted height and corresponding
# header hash obtained from a trusted source, and a period during which validators can be trusted.
#
# For Cosmos SDK-based chains, trust_period should usually be about 2/3 of the unbonding time (~2
# weeks) during which they can be financially punished (slashed) for misbehavior.
rpc_servers = "http://<IP_ADDRESS>:80,http://<IP_ADDRESS>:80"
trust_height = 1569274
trust_hash = "59E5CBC5E2B0B3E251FC3667F4E8CC2DFB7ED12F1C041252E2169022CAC3A899"
trust_period = "168h0m0s"
# Time to spend discovering snapshots before initiating a restore.
discovery_time = "15s"
# Temporary directory for state sync snapshot chunks, defaults to the OS tempdir (typically /tmp).
# Will create a new, randomly named directory within, and remove it when done.
temp_dir = ""
# The timeout duration before re-requesting a chunk, possibly from a different
# peer (default: 1 minute).
chunk_request_timeout = "10s"
# The number of concurrent chunk fetchers to run (default: 1).
chunk_fetchers = "4"
However, the node does not sync. Instead, I see these logs:
6:08PM INF starting node with ABCI Tendermint in-process module=server
6:08PM INF service start impl=multiAppConn module=proxy msg={}
6:08PM INF service start connection=query impl=localClient module=abci-client msg={}
6:08PM INF service start connection=snapshot impl=localClient module=abci-client msg={}
6:08PM INF service start connection=mempool impl=localClient module=abci-client msg={}
6:08PM INF service start connection=consensus impl=localClient module=abci-client msg={}
6:08PM INF service start impl=EventBus module=events msg={}
6:08PM INF service start impl=PubSub module=pubsub msg={}
6:08PM INF service start impl=IndexerService module=txindex msg={}
6:08PM INF Version info abci=1.0.0 block=11 commit_hash= module=server p2p=8 tendermint_version=0.37.1
6:08PM INF This node is not a validator addr=49BF5A9A0C4117F5EB1E51AFFB59755D8371A8FA module=consensus pubKey=uEUPyozNqgYUQsSL9KWHJ4aas3YkiEczknKJKB08jpQ=
6:08PM INF P2P Node ID ID=0b8f89eb909b9098fe19af1d4d61e3e257d31c5a file=/home/ashish/.ssc/config/node_key.json module=p2p
6:08PM INF Adding persistent peers addrs=["e7f5e5327a8298eb04c29c4115ccf2d6a05ec732@<IP_ADDRESS>:26656","3abc213ec08ece180e6fa1443226689ecc4b7749@<IP_ADDRESS>:26656","239b182bab3252c26fc116bf00301f30f1dea01e@<IP_ADDRESS>:26656","8d936668e433e9e50cd4e60da218a9fe81950d3f@<IP_ADDRESS>:26656","65b64dc2d28e0116da22582cb947ec7bc8c91173@<IP_ADDRESS>:26656"] module=p2p
6:08PM INF Adding unconditional peer ids ids=[] module=p2p
6:08PM INF Add our address to book addr={"id":"0b8f89eb909b9098fe19af1d4d61e3e257d31c5a","ip":"<IP_ADDRESS>","port":26656} book=/home/ashish/.ssc/config/addrbook.json module=p2p
6:08PM INF service start impl=Node module=server msg={}
6:08PM INF Starting pprof server laddr=localhost:6060 module=server
6:08PM INF serve module=rpc-server msg={}
6:08PM INF service start impl="P2P Switch" module=p2p msg={}
6:08PM INF service start impl=Reactor module=blockchain msg={}
6:08PM INF service start impl=ConsensusReactor module=consensus msg={}
6:08PM INF Reactor module=consensus waitSync=true
6:08PM INF service start impl=Evidence module=evidence msg={}
6:08PM INF service start impl=StateSync module=statesync msg={}
6:08PM INF service start impl=PEX module=pex msg={}
6:08PM INF service start book=/home/ashish/.ssc/config/addrbook.json impl=AddrBook module=p2p msg={}
6:08PM INF Saving AddrBook to file book=/home/ashish/.ssc/config/addrbook.json module=p2p size=30
6:08PM INF Ensure peers module=pex numDialing=0 numInPeers=0 numOutPeers=0 numToDial=10
6:08PM INF Starting state sync module=statesync
6:08PM INF Downloading trusted light block using options module=light
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} 8d936668e433e9e50cd4e60da218a9fe81950d3f out}" module=p2p msg={} peer={"id":"8d936668e433e9e50cd4e60da218a9fe81950d3f","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"8d936668e433e9e50cd4e60da218a9fe81950d3f","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} fa64dcb9be1733fe27e932b3d5da1e685be6906a out}" module=p2p msg={} peer={"id":"fa64dcb9be1733fe27e932b3d5da1e685be6906a","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"fa64dcb9be1733fe27e932b3d5da1e685be6906a","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} 3abc213ec08ece180e6fa1443226689ecc4b7749 out}" module=p2p msg={} peer={"id":"3abc213ec08ece180e6fa1443226689ecc4b7749","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"3abc213ec08ece180e6fa1443226689ecc4b7749","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} e3f30a584c09a2a599ab58fbe8fc2958faf6c351 out}" module=p2p msg={} peer={"id":"e3f30a584c09a2a599ab58fbe8fc2958faf6c351","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"e3f30a584c09a2a599ab58fbe8fc2958faf6c351","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF sync any module=statesync msg={}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} 82417f944fbf680042de3c4acf33aed51b34e8ec out}" module=p2p msg={} peer={"id":"82417f944fbf680042de3c4acf33aed51b34e8ec","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"82417f944fbf680042de3c4acf33aed51b34e8ec","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} d9dabf20d621a021aac1477b5a714cd5a8d8a3b5 out}" module=p2p msg={} peer={"id":"d9dabf20d621a021aac1477b5a714cd5a8d8a3b5","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"d9dabf20d621a021aac1477b5a714cd5a8d8a3b5","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} e7f5e5327a8298eb04c29c4115ccf2d6a05ec732 out}" module=p2p msg={} peer={"id":"e7f5e5327a8298eb04c29c4115ccf2d6a05ec732","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"e7f5e5327a8298eb04c29c4115ccf2d6a05ec732","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} 65b64dc2d28e0116da22582cb947ec7bc8c91173 out}" module=p2p msg={} peer={"id":"65b64dc2d28e0116da22582cb947ec7bc8c91173","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"65b64dc2d28e0116da22582cb947ec7bc8c91173","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl="Peer{MConn{<IP_ADDRESS>:26656} 239b182bab3252c26fc116bf00301f30f1dea01e out}" module=p2p msg={} peer={"id":"239b182bab3252c26fc116bf00301f30f1dea01e","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF service start impl=MConn{<IP_ADDRESS>:26656} module=p2p msg={} peer={"id":"239b182bab3252c26fc116bf00301f30f1dea01e","ip":"<IP_ADDRESS>","port":26656}
6:08PM INF sync any module=statesync msg={}
6:08PM INF Ensure peers module=pex numDialing=0 numInPeers=0 numOutPeers=9 numToDial=1
6:08PM INF We need more addresses. Sending pexRequest to random peer module=pex peer={"Data":{},"Logger":{"Logger":{}}}
6:08PM INF sync any module=statesync msg={}
I check the status on my local node and it shows me this:
{"NodeInfo":{"protocol_version":{"p2p":"8","block":"11","app":"0"},"id":"0b8f89eb909b9098fe19af1d4d61e3e257d31c5a","listen_addr":"tcp://<IP_ADDRESS>:26656","network":"ssc-testnet-1","version":"0.37.1","channels":"40202122233038606100","moniker":"state-sync-node","other":{"tx_index":"on","rpc_address":"tcp://<IP_ADDRESS>:26657"}},"SyncInfo":{"latest_block_hash":"","latest_app_hash":"","latest_block_height":"0","latest_block_time":"1970-01-01T00:00:00Z","earliest_block_hash":"","earliest_app_hash":"","earliest_block_height":"0","earliest_block_time":"1970-01-01T00:00:00Z","catching_up":true},"ValidatorInfo":{"Address":"49BF5A9A0C4117F5EB1E51AFFB59755D8371A8FA","PubKey":{"type":"tendermint/PubKeyEd25519","value":"uEUPyozNqgYUQsSL9KWHJ4aas3YkiEczknKJKB08jpQ="},"VotingPower":"0"}}
Anything in the debug logs --log_level="*:debug"? https://docs.cometbft.com/main/explanation/core/running-in-production#logging
Anything in the debug logs --log_level="*:debug"? https://docs.cometbft.com/main/explanation/core/running-in-production#logging
Sorry just saw this. I will revert back today with the logs.
@melekes please see the debug log from today:
sscd.log
Hi @melekes or anyone from the CometBFT team,
Any thoughts or way forward on this?
Hey guys, any updates on this? I realized I also needed seed nodes and I have added those to my node's config.toml and tried running this. Still no luck - same issue.
@hvanz could you please take a look at this issue (as a part of the state-sync sprint)?
I looked at the log; it seems that Comet requests the list of snapshots from its peers, but it never gets a response. Is it possible that the problem is in the implementation of ListSnapshots in the application?
I looked at the log; it seems that Comet requests the list of snapshots from its peers, but it never gets a response. Is it possible that the problem is in the implementation of ListSnapshots in the application?
@hvanz It is a vanilla Cosmos SDK application and in a public repo. Feel free to look at what we are doing: https://github.com/sagaxyz/ssc/blob/c4a5823177409f4ee0ea4f99f0a7cf0c412668a7/cmd/sscd/cmd/root.go#L228
Hi @melekes @hvanz I figured out the issue after @hvanz indicated it was not finding a snapshot. We have persistent peers values and rpc nodes values. The RPC nodes are the ones that have the state-sync snapshots and our persistent peers listed in the config.toml do not have state-sync snapshots.
My local node is now running and syncing with a state-sync snaphot.
I did not find this requirement for the persistent peers to have the state-sync snapshot nodes listed. I only knew of the rpc_servers value, which was correctly set. I may have missed this in the docs but in case it is not there, please add the requirement to also have the persistent_peers include the nodes that are serving the state-sync snapshots.
@ashishchandr70 glad you've figured it out 👍 we will check the docs, thanks
Hi @ashishchandr70, I'm glad that you found the problem. Indeed, the snapshots are requested from all known peers, including the persistent peers. It's mentioned in the code but probably the documentation should be more explicit about it. https://github.com/cometbft/cometbft/blob/v0.37.1/statesync/reactor.go#L272-L278
@hvanz could you please make it clear in the docs?
|
2025-04-01T06:38:15.190493
| 2023-10-06T08:19:28
|
1929668614
|
{
"authors": [
"adizere",
"cason"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4842",
"repo": "cometbft/cometbft",
"url": "https://github.com/cometbft/cometbft/pull/1438"
}
|
gharchive/pull-request
|
grpc: Add base gRPC server with version service to v0.37.x-experimental
Addresses #1420.
Refer to the companion PR for v0.38 for additional details: #1437
PR checklist
[x] Tests written/updated
[x] Changelog entry added in .changelog (we use unclog to manage our changelog)
[ ] Updated relevant documentation (docs/ or spec/) and code comments
Closing until we get signals that this is needed in v0.37.
|
2025-04-01T06:38:15.193640
| 2023-03-24T13:35:01
|
1639406607
|
{
"authors": [
"jmalicevic",
"sergio-mena"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4843",
"repo": "cometbft/cometbft",
"url": "https://github.com/cometbft/cometbft/pull/579"
}
|
gharchive/pull-request
|
rpc: tweaked the block search test
Improved the BlockSearch test within the RPC to make sure it matches against a value if the block events are generated.
The previous version had a query that would never match and this way it is harder to test whether indexing was performed correctly.
PR checklist
[x ] Tests written/updated
[ ] Changelog entry added in .changelog (we use unclog to manage our changelog)
[ ] Updated relevant documentation (docs/ or spec/) and code comments
Should this be backported?
I added the backport labels, confirmed it works on all branches, but will need some tweaking for 0.34 as it has this match_events keyword. I'll take care of it.
|
2025-04-01T06:38:15.205729
| 2023-09-04T08:01:15
|
1879730966
|
{
"authors": [
"cason",
"josef-widder",
"sergio-mena"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4844",
"repo": "cometbft/knowledge-base",
"url": "https://github.com/cometbft/knowledge-base/pull/14"
}
|
gharchive/pull-request
|
Addressing comet/comet#1174. First version
As per title.
rendered
Once you're OK with the text, I'm planning to adapt the structure so that it is an RFC in the Comet repo.
There are two "TODOs" in the "Eventual requirements" section. Please let me know what you think.
Thanks folks for all your comments 🙏. I'll go through them tomorrow 👀
Notice that the rendered version should point to the latest version on the branch, namely: https://github.com/cometbft/knowledge-base/blob/sergio/addressing-1174/protocols/abci/addressing-1174.md
As a generic comment, I think this is a very nice discussion and problem to address. Some practical terms in think we should discuss and find more precise definitions:
The external validity definition needs some improvement; we should look how validity is handled by other protocols. In general, the property of BFT consensus is that a value proposed by a correct value is eventually decided. For values proposed by Byzantine agents, there are multiple definitions and ways to deal with
The changes in the consensus protocol are interesting. I wonder whether we can replace locked value/round by valid value/round on the properties. We need in fact to consider the corner cases when they differ. But they can only differ by having valid round/value more updated (fresher) than locked round/value.
As a generic and simplistic definition: a valid value is a value that I would propose; a locked value is a valued that I have accepted (prevoted for)
Regarding the pseudo-code, we have to recall that the introduction of the valid value is posterior to the writing of this pseudo-code. That is why it is a little confusing, as the valid part was introduced afterwards in order to handle with hidden locks and ensure progress after GST.
An idea that I had, I would like some feedback on it, is that we do not remove the valid(v) calls from the algorithm, in none of the clauses.
We should be able, instead, of writing a valid(v) method that does consider:
Whether we have already validated v, in which case we might skip the validation cost and just return the previous value produced by the external validation (application)
Whether we observed a quorum of processes that have validated v, in the form of 2f+1 Prevote(id(v); this allows us, again, to avoid the external validation by the application
Whether we observed a a quorum of processes that have accepted v, in the form of 2f+1 Precommit(id(v)
The main point that I have here is that a process must validate a proposed value at least once. The question is whether a rejection of a proposal block by a non-deterministic external validation procedure should prevent us from committing (or enable the commitment) of a block. The validation check performed by the blockchain (i.e., is the block a valid block?) should be performed at least once for every committed block.
In the spirit of PBTS, notice that the change introduced there was to create two validation methods, the valid(v) that remained untouched, and an additional non-deterministic timely(v) predicate. This second predicate was only evaluated when a block was received for the first time, as it was a timing predicate. In the case of re-proposed blocks, the fact that $2f + 1$ processes have considered the block timely enabled us to drop the timely check in this case.
An idea that I had, I would like some feedback on it, is that we do not remove the valid(v) calls from the algorithm, in none of the clauses.
I guess this is related to the separation of valid as a mathematical function (where it doesn't matter how often it is called) and Process_Proposal (where it might matter how often it is called if it uses oracle data). The question is whether we want to capture the latter point in the pseudo code...
I think at a very early stage, Dev has written a pseudo code variant of ABCI++ with Process_Proposal. The question is whether we should formalize it in this way, and thus eliminate the "schisma" between valid and Process_Proposal?
Another question that we need to address is how to handle the recovery of a node.
The assumption that valid(v) is a deterministic method enables us to employ this method both in regular operation and in recovery mode. So, if a node during regular operation has deemed a block valid, therefore it has issued a Prevote for the block, when this situation arises again in recovery mode, the validation should produce the same output. Otherwise, if the block is reject in this different context, the node will Prevote nil, which is an equivocation.
This same question was raised for PBTS. The solution proposed (but not implemented) is to persist in the consensus WAL the information needed to evaluate the timely predicate. Namely, we have to persist the receive time of a proposal, which is the non-deterministic part of the timely evaluation. The method, by itself, is deterministic, so that we need only to ensure that it is invoked with the same parameters both in regular and recovery mode.
How we can handle this situation when ProcessProposal is not deterministic? The first solution that came to mind is to persist in the WAL the output of this validation method when it is invoked during regular operation. The invocation of this method therefore becomes a non-deterministic external event processed by the consensus protocol. Since we probably cannot replay the exactly same (application and timing) context used during regular operation, we don't have much options than to store the result of this operation in disk. Once in recovery mode, we should be able to identify that we have the output of this validation and therefore must not perform it again, with the risk of producing a different behavior at consensus level (i.e., equivocating).
Addressed outstanding comments. I think this is ready to land as an RFC PR on CometBFT repo. If there are further comments, they can be added there. Merging this
|
2025-04-01T06:38:15.256839
| 2024-08-24T07:41:58
|
2484331149
|
{
"authors": [
"lordwelch",
"protofolius"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4845",
"repo": "comictagger/comictagger",
"url": "https://github.com/comictagger/comictagger/issues/672"
}
|
gharchive/issue
|
complicated parser and ' or + character
filename
Airfiles #04 The 'Big Show' 2.cbz
turns title into
The' Big Show' 2
filename
Army #11 Operation 'Z'.cbz
turns title into
Operation' Z'
filename
De Dwaas Van De Koning #009 Het Testament Van d'Artagnan.cbz
turns title into
Het Testament Van d' Artagnan
notice the added space after the '
I am using the complicated parser through command line , but the same behaviour is seen in the GUI .
Using version 1.6.0a21.dev0
The same happens with + signs .
filename
De Man of Steel #01 The Invasion + Power of The Monster.cbz
turns title into
The Invasion+ Power of The Monster
When a title start with a ' character , it drops it completely .
For example
Atari Force #09 'Als Uw Oog U Tot Last Is...'.cbz
turns title into
Als Uw Oog U Tot Last Is'
drops the first ' completely (and it also drops the ... for some reason)
that's because quotes shouldn't be in filenames 😝. I'll see if I can get to this next week
If you need a Comic series to test with , this would be a good example for the ' sign :
https://www.stripinfo.be/reeks/index/1847_De_psy
|
2025-04-01T06:38:15.269782
| 2020-04-26T02:40:37
|
606901907
|
{
"authors": [
"AhadCove",
"codename224",
"jyoung8607",
"pd0wm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4846",
"repo": "commaai/openpilot",
"url": "https://github.com/commaai/openpilot/issues/1423"
}
|
gharchive/issue
|
Joystickd readme in tools outdated
Joystickd in tools does not seem to function, and it's readme instructions are out of date.
When running all three scripts with no visible console errors and all dependencies working correctly, joystickd.py does not move the steering wheel.
The readme instructions are also out of date. For example, you do not need to manually set ALL_OUTPUT anymore.
Everything should still work, I used it pretty recently. I updated the readme with some new instructions. Can you get it to work with those instructions?
Another thing that will stop joystick use (and MOCK boardd, and a lot of the Panda Python example scripts) is running firmware built for EON, including the default/signed shipped FW from Comma. In short, a Panda flashed for actual driving with openpilot won't work for any of these tools.
Firmware built with the EON option expects a keepalive from OP boardd, and the missing keepalive kicks Panda back to NOOUTPUT within a couple seconds after the tool sets an output mode.
It's very easy to run into this, and hard for a newbie to figure out because the failure is totally silent, and it's something of an annoyance to reflash Panda if your usual development environment is virtualized.
I ran into this the hard way recently, but had not submitted a PR to fix it because I'm not sure of a workable approach that Comma would find acceptable. I don't like having to flash back and forth, and I'd especially like to be able to use the Panda tools from a vehicle-installed EON/C2. For my own development purposes, I hacked around it by setting the heartbeat timeout to an entire day.
Alternatives are to add a USB command to disable (or lengthen) keepalives, or to augment all the various development tools to send the heartbeat commands periodically. Or, even if no automatic workaround is acceptable, we could at least have the dev tools detect EON firmware and bail out with an informative notice.
I tried to get joystickd to work as well for some time to no avail. But what you mentioned @jyoung8607 sounds pretty much the same issue I was having, it was silently failing, but the code was still running and looked as if it was working, but nothing would happen.
If you have a simple guide to what you did to get past this it would be extremebly helpful. Did you create your own panda firmware with the heartbeat or eon firmware?
I think if you follow Willem's updated instructions, it's more likely to work for officially supported cars since it's using the new boardd that emits keepalives. I have not tested this personally with an EON or Comma signed build.
For the old instructions using boardd_old.py, with the default Panda firmware, I think you would have trouble with missing keepalives on EON builds, and also you can't go into ALL_OUTPUT mode without DEBUG firmware built from source. You'd also have trouble if your car's safety mode is in the ALLOW_DEBUG group as not yet officially supported.
You should retry with Willem's updated instructions. If that doesn't work, you should build the Panda firmware yourself from source, ideally on Ubuntu bare metal (virtualization makes it difficult to carry out the reflash step) so all the DEBUG stuff is enabled for you.
|
2025-04-01T06:38:15.272623
| 2020-05-18T15:45:48
|
620314034
|
{
"authors": [
"zoukyle"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4847",
"repo": "commaai/openpilot",
"url": "https://github.com/commaai/openpilot/issues/1534"
}
|
gharchive/issue
|
Couldn't make the black panda to turn the wheel by sending the STEERING_LKA CAN command
The wheel can turn a little, most of the time not even noticeable. After debugging, I found that increasing the torque cannot exceed a limit (between 350 and 400) and the steering angle only changes a little.
In addition, the EPS_STATUS LKA_STATE becomes 9 after the torque increases beyond the limit. I also found if the torque is below the limit, the LKA_STATE flickers between 1 and 5 quickly, which I don't think it's normal.
After installing the black panda, should the car itself still send the STEERING_LKA message to the CAN bus? It seems that when I print out the logcan messages, the car itself is still sending bus: 000 {'CHECKSUM': 78.0, 'LKA_STATE': 0.0, 'STEER_REQUEST': 0.0, 'STEER_TORQUE_CMD': 0.0}. Is this expected?
My theory is that the STEERING_LKA messages sent from the car itself interferes with the messages sent from openpilot. Not sure if that's case.
After debugging, I found out the issue:
Can1 (bus number 0) should not get the LKA 740 messages from the camera (FRC). The messages from the camera will cause conflict with the 740 messages that I send by using the openpilot joystick controller.
|
2025-04-01T06:38:15.274424
| 2020-08-26T11:50:05
|
686255209
|
{
"authors": [
"kyeli2",
"pd0wm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4848",
"repo": "commaai/openpilot",
"url": "https://github.com/commaai/openpilot/pull/2085"
}
|
gharchive/pull-request
|
Systemd logcatd
I'm in need for sone help on my Honda Insight. I have openpilot and sent back because I asked about brightness. They sent a new one and now it says in is unsupported. Can you help?
What do I do with this information?
I load branch "master" and the camera dent go on.
|
2025-04-01T06:38:15.280262
| 2016-07-18T10:39:06
|
166065377
|
{
"authors": [
"MGA-dotSource",
"floriansattler"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4849",
"repo": "commercetools/commercetools-sunrise-java-payment",
"url": "https://github.com/commercetools/commercetools-sunrise-java-payment/issues/5"
}
|
gharchive/issue
|
Country specific config for payment methods
Config for payments needs to be done in a way that you are able to show different payment methods in the checkout based on the country selected in your billing address.
This is a bigger task.
First question: is (or should) the abstract framework be able to decide about allowed / not allowed countries?
If yes, then what are the rules?
If it is a shop responsibility then there is already a build in feature, that allows the shop to pass a filter function into the "getMethods" - API call that is executed every time where any kind of filtering is possible.
Filtering itself has to be discussed in more detail and is not in scope for the first iteration I think.
|
2025-04-01T06:38:15.319156
| 2024-12-12T10:39:30
|
2735521062
|
{
"authors": [
"chreekat",
"juhp"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4850",
"repo": "commercialhaskell/stackage-server",
"url": "https://github.com/commercialhaskell/stackage-server/pull/338"
}
|
gharchive/pull-request
|
a fix and a url update
fix dates in blog archive list to not be for current displayed post
update url to stackage readme section on adding package
bump to final lts-22.43
Nice, thanks. :) I'll work on getting this deployed, which is still a bit of a bad process.
Thanks
Finally deployed now. :)
|
2025-04-01T06:38:15.349273
| 2023-11-11T01:44:55
|
1988640837
|
{
"authors": [
"commondatageek"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4851",
"repo": "commondatageek/mark",
"url": "https://github.com/commondatageek/mark/issues/36"
}
|
gharchive/issue
|
Don't error if there is no $HOME/.bookmarks.jsonl
Instead, give a helpful suggestion:
try mark add {url} !
Fixed in #45
|
2025-04-01T06:38:15.414164
| 2016-03-01T04:01:09
|
137461605
|
{
"authors": [
"chrisdebian",
"misaochan",
"nicolas-raoul"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4852",
"repo": "commons-app/apps-android-commons",
"url": "https://github.com/commons-app/apps-android-commons/issues/76"
}
|
gharchive/issue
|
Headers to distinguish where suggestions come from
How about placing small headers to show why suggestions are made?
Prototype:
--- Nearby ---
Trafalgar Square
Buckingham Palace
--- Guessed from name ---
Demonstration
Climate change
--- Recent ---
Roti prata
Rendang
When I take a picture, I know already what type of suggestion will be correct. For instance if the last upload was of the same thing, I know I must select everything in "Recent". If I am in a famous place but photographing something unrelated to this place (for instance a vehicle) then I can ignore all "Nearby" suggestions. Etc. Having sections will thus make selection faster.
It could also allow making the suggestions list a bit longer without it feeling too long?
Hm, this sounds like a great suggestion, but I think I will have to pick between this and #70 for the grant proposal, due to the time limitations. Which do you think would be more important?
Difficult choice haha. It does not sound as difficult as anonymization, and will probably be more useful to most people.
Okay, this one it is then! :)
How do we want the headers to look like? Should they actually take up slots in the ListView like all the categories, just with the '--' to make it obvious that they are headers?
Is there an easy to use component that provide this? Either a ListView with a special setting, or a child class of ListView, or a totally different component.
I was envisioning something like this, maybe with smaller and less eye-catching headers:
http://stacktips.com/tutorials/android/listview-with-section-header-in-android
Hm. This is turning out to be much more complicated than I expected.
I've tried to apply this tutorial (and other similar ones) to our code, but the main issue is that all the solutions that I've found involve adding the section headers to the ArrayList that is used as the data source in the adapter. So now our data source contains not just CategoryItems but also header items.
This could be workable, but in CategorizationFragment.java most of the existing code assumes that every item in the data source is a Category, and operates on it accordingly. Also most of the existing methods rely on using the index of an element in the ArrayList for category selection, assignment, etc. It feels rather messy to have headers be part of that array, and to have to keep checking for getItemViewType(position) whenever we want to perform any operations on categories? I tried this method anyway at https://github.com/misaochan/apps-android-commons/commits/header-list-2 , but stopped after it became unwieldy.
I wonder if there might be a better way to handle this, given that it is only a UI change, so we probably shouldn't be introducing all these extra elements into the data source?
Is there no way to have each row contain an object, and use this object
rather than using the row's index?
On Sun, Nov 6, 2016 at 3:38 PM, Josephine Lim<EMAIL_ADDRESS>wrote:
Hm. This is turning out to be much more complicated than I expected.
I've tried to apply this tutorial (and other similar ones) to our code,
but the main issue is that all the solutions that I've found involve adding
the section headers to the ArrayList that is used as the data source in the
adapter. So now our data source contains not just CategoryItems but also
header items.
This could be workable, but in CategorizationFragment.java most of the
existing code assumes that every item in the data source is a Category, and
operates on it accordingly. Also most of the existing methods rely on using
the index of an element in the ArrayList for category selection,
assignment, etc. It feels rather messy to have headers be part of that
array, and to have to keep checking for getItemViewType(position)
whenever we want to perform any operations on categories? I tried this
method anyway at https://github.com/misaochan/
apps-android-commons/commits/header-list-2 , but stopped after it became
unwieldy.
I wonder if there might be a better way to handle this, given that it is
only a UI change, so we probably shouldn't be introducing all these extra
elements into the data source?
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/commons-app/apps-android-commons/issues/76#issuecomment-258663552,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAGFBvfAg7df6HaWkg-KFUwW1OdqsnFsks5q7XYDgaJpZM4HmH_N
.
I'm not sure, but it would involve an almost complete refactor of CategorizationFragment.
Alternatively, I was thinking of trying a 3rd party library: https://github.com/commonsguy/cwac-merge
The 3rd party libraries I mentioned are on a Apache 2.0 license. Should be okay to include them, right?
Apache 2.0 is OK, yes.
The 3rd party library works (as seen in https://github.com/misaochan/apps-android-commons/tree/header-list-cwac-merge , headers are produced for GPS, recent, and title cats), but further issues have been discovered along the way.
The main issue is that in order to get this to work, I have needed to create a separate adapter for the empty search field (which displays automatically-generated GPS, recent, and title cats along with their headers), and for the non-empty search field (which displays the results of a manual category search with no headers).
In the app currently, selected categories are aggregated and displayed at the top of the list, from any type of category suggestion (manual or automatic). This is really useful and something I don't want to lose. But with two separate adapters, I can't seem to get this to happen anymore - selected categories remain selected but are only displayed in their respective adapter (automatic cats that are selected only display when search field is empty, manual cats only display when search field is not empty).
At this stage I think the amount of work needed for me to complete this task is exceeding the benefits that it would bring (headers would be really nice but I don't think it is a dealbreaker for most users). It's entirely possible that I'm missing something simple, so I will leave this issue and my associated branch open, but will move on to the next task for now.
Would you mind posting a screenshot of what it looks like with
header-list-cwac-merge ?
Thanks!
On Mon, Nov 14, 2016 at 2:51 PM, Josephine Lim<EMAIL_ADDRESS>wrote:
The 3rd party library works (as seen in https://github.com/misaochan/
apps-android-commons/tree/header-list-cwac-merge , headers are produced
for GPS, recent, and title cats), but further issues have been discovered
along the way.
The main issue is that in order to get this to work, I have needed to
create a separate adapter for the empty search field (which displays
automatically-generated GPS, recent, and title cats along with their
headers), and for the non-empty search field (which displays the results of
a manual category search with no headers).
In the app currently, selected categories are aggregated and displayed at
the top of the list, from any type of category suggestion (manual or
automatic). This is really useful and something I don't want to lose. But
with two separate adapters, I can't seem to get this to happen anymore -
selected categories remain selected but are only displayed in their
respective adapter (automatic cats that are selected only display when
search field is empty, manual cats only display when search field is not
empty).
At this stage I think the amount of work needed for me to complete this
task is exceeding the benefits that it would bring (headers would be really
nice but I don't think it is a dealbreaker for most users). It's entirely
possible that I'm missing something simple, so I will leave this issue and
my associated branch open, but will move on to the next task for now.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/commons-app/apps-android-commons/issues/76#issuecomment-260253875,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAGFBgBwj5F_QtmXxRPiCSh23JlFzTfKks5q9_bdgaJpZM4HmH_N
.
Sure. This is of the automatic categories (I hadn't hooked it up to the filterYears() method yet, so years are still being displayed):
And this is of the manual categories:
This looks like a superb idea, and the screenshots above are really helpful. Is this still something that needs to be implemented?
Chris.
Yes.
Could we add this to a target milestone, then; it's been open for eight years? It looks like people support it as an idea, so if we have a target release, it may attract volunteers who want to write the code.
Just an idea.
This isn't meant to sound harsh, but it feels like good ideas Luke this, should at least be on the 'To Do' list.
Chris
|
2025-04-01T06:38:15.419116
| 2022-04-23T03:35:38
|
1213181883
|
{
"authors": [
"AlexMahlon",
"nicolas-raoul"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4853",
"repo": "commons-app/apps-android-commons",
"url": "https://github.com/commons-app/apps-android-commons/pull/4941"
}
|
gharchive/pull-request
|
Fixes #4934 enforced Wikimedia character blacklisting on captions and…
Description (required)
Originally the caption and description TextViews would accept any characters. It was possible to upload files with banned characters. Now, an input filter is used to remove these banned characters.
Fixes #4934
What changes did you make and why?
A new InputFilter class was applied to the caption and description TextViews. This filter checks for any illegal characters and removes them. The set of illegal characters can be found here: blacklist.
Tests performed (required)
Tested ProdDebug on Pixel 6 with API level 31.
Also included unit tests for the new InputFilter class.
I just tested, it works great and indeed fixes the bug.
Hi Alex!
Did you get a chance apply the minor modifications?
If you have no time, I can also do it for you, like you want :-)
Hi Nicolas.
Sorry for the delay, I had to do some traveling.
The new PR can be found here: #4955.
No problem, I hope you had a nice trip!
Tip: you can update an existing pull request, simply by pushing new commits to it.
|
2025-04-01T06:38:15.422476
| 2022-05-22T12:34:33
|
1244241425
|
{
"authors": [
"neslihanturan",
"nicolas-raoul"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4854",
"repo": "commons-app/apps-android-commons",
"url": "https://github.com/commons-app/apps-android-commons/pull/4969"
}
|
gharchive/pull-request
|
[WIP] Explore nearby pictures test
Description (required)
Fixes #INSERT_ISSUE_NUMBER_HERE
What changes did you make and why?
Tests performed (required)
Tested {build variant, e.g. ProdDebug} on {name of device or emulator} with API level {API level}.
Screenshots (for UI changes only)
Need help? See https://support.google.com/android/answer/9075928
Note: Please ensure that you have read CONTRIBUTING.md if this is your first pull request.
This was successfully implemented in another PR I believe. :-)
|
2025-04-01T06:38:15.429329
| 2021-09-01T05:52:39
|
984691369
|
{
"authors": [
"dealako",
"nickmango",
"thakurveerendras"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4855",
"repo": "communitybridge/easycla",
"url": "https://github.com/communitybridge/easycla/issues/3230"
}
|
gharchive/issue
|
"Cannot read property 'length' of undefined" error appears when open easyCla page at PCC
Steps :
Login admin portal
Search Project which have cla enabled
Open easyCla page
Actual Results : "Cannot read property 'length' of undefined" error appears & page is inaccessible
ERROR TypeError: Cannot read property 'length' of undefined
at ActivityLogComponent_Template (activity-log.component.html:2)
at executeTemplate (core.js:9614)
at refreshView (core.js:9480)
at refreshComponent (core.js:10651)
at refreshChildComponents (core.js:9277)
at refreshView (core.js:9530)
at refreshEmbeddedViews (core.js:10605)
at refreshView (core.js:9504)
at refreshEmbeddedViews (core.js:10605)
at refreshView (core.js:9504)
https://images.zenhubusercontent.com/194341141/d542169a-41d2-40e1-b3e8-d9834843c814/easycla_error.mp4
@thakurveerendras the error is thrown on activity-log.component.html so looks like a PCC issue on the frontend
Thanks @nickmango ,
Added frontend ticket for above case (https://jira.linuxfoundation.org/browse/PCC-1591).
Closing as this appears to be a front-end issue in the PCC. Thanks!
|
2025-04-01T06:38:15.441289
| 2016-06-15T18:06:37
|
160486802
|
{
"authors": [
"EyesOnlyNet",
"drixie"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4856",
"repo": "compact/angular-bootstrap-lightbox",
"url": "https://github.com/compact/angular-bootstrap-lightbox/issues/65"
}
|
gharchive/issue
|
Activate Lightbox with custom HTML
Let's say I don't want to use img, and I can to activate the lightbox on , how do I achieve this?
I actually tried it but console.log gave "Invalid Images"
Thanks in advance!
If you want to achieve a simple modal, without connection to any images, you should use the normal UI-Modal (https://angular-ui.github.io/bootstrap/#/modal).
|
2025-04-01T06:38:15.446506
| 2023-10-31T19:19:42
|
1971153495
|
{
"authors": [
"Zain-Abbas1",
"markpit"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4857",
"repo": "companieshouse/overseas-entities-web",
"url": "https://github.com/companieshouse/overseas-entities-web/pull/1181"
}
|
gharchive/pull-request
|
ROE-2405 Trusts associated with OE
JIRA link
https://companieshouse.atlassian.net/browse/ROE-2405
Change description
Update index files to include new url with params
Update html for change link with params
Update trust involved page to redirect to next page with url params
Update add.trust to include ActiveSubmissionBasePath in html template
Work checklist
[x] Tests added where applicable
[ ] UI changes meet accessibility criteria
Merge instructions
We are committed to keeping commit history clean, consistent and linear. A BREAKING CHANGE can be part of commits of any type,
types other than fix: and feat: are allowed, for example build:, chore:, ci:, docs:, style:, refactor:, perf:, test:, and others,
footers other than BREAKING CHANGE: <description> may be provided.
looks ok to me once dans's comments resolved
|
2025-04-01T06:38:15.477119
| 2024-10-14T10:51:29
|
2585642279
|
{
"authors": [
"gaurovgiri"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4858",
"repo": "computerclubkec/computerclubkec.github.io",
"url": "https://github.com/computerclubkec/computerclubkec.github.io/pull/97"
}
|
gharchive/pull-request
|
feat: add upcoming events page that displays all the upcoming events
This PR resolves #92 by displaying all the upcoming events in a single dedicated page.
Output:
It's look better. I have done for #98 as well.
|
2025-04-01T06:38:15.808285
| 2022-02-02T15:05:38
|
1122039850
|
{
"authors": [
"kswenson"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4859",
"repo": "concord-consortium/collaborative-learning",
"url": "https://github.com/concord-consortium/collaborative-learning/pull/1187"
}
|
gharchive/pull-request
|
Fix promise-handling bug in creation of planning documents
[#181124054]
Fixes a copy/paste/logic bug in planning document promise-handling which likely explains the occurrence of extraneous planning documents in some circumstances.
@scytacki I also added tests that would have caught the bug, officially closing the barn door. Adding the tests required setting up offline mocking of firebase that should prove useful for writing additional tests down the road.
The database mocking is complex. I could see it being hard to maintain. It would probably fail in a weird way if some code under test makes an unexpected firebase call. But lets see how it goes. :)
It's coupled tightly to the code under test, which certainly makes it more fragile, but is also why I restricted the mocking to within each individual test. At least this way when you're restructuring some code the tests of that specific code are likely to break, but it won't require updates to a shared mock that then cascades to other failing tests. In this particular case, the important point is that it tests the promise-handling logic, which was the point of the exercise.
|
2025-04-01T06:38:15.811886
| 2023-07-18T12:23:55
|
1809820433
|
{
"authors": [
"bgoldowsky",
"scytacki"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4860",
"repo": "concord-consortium/collaborative-learning",
"url": "https://github.com/concord-consortium/collaborative-learning/pull/1820"
}
|
gharchive/pull-request
|
Refactor how dragging to move objects works.
Simplifies how we track the temporary position of objects on the drawing canvas while they are being dragged to move them.
Previously this was done by holding an "objectsBeingDragged" list in state, hiding these objects in the render phase, and creating clones of these objects outside of the model with altered x and y coordinates.
The new code uses volatile dragX and dragY coordintes that are part of the object model and used by the normal render process. Having these be volatile prevents each step of the drag from adding an undo step.
This concept is borrowed from https://github.com/concord-consortium/quantity-playground/blob/main/src/diagram/models/dq-node.ts#L27
The UX looks good. I think I'm forgetting some of the tricky dragging and selection cases.
If you want to test the variable chip object you can use the example unit. Then add the second node-arc tile. The rollover of the button on the toolbar is "Diagram". Then add the drawing tool. Now you should see a V= tool bar button on the drawing tool. Clicking this will add a variable chip to the drawing tool.
Here is the link to test it on the deployed branch:
https://collaborative-learning.concord.org/branch/refactor-draw-object-drag/?appMode=dev&unit=example
Another nice side effect of this, change is that if you open the same document on the left and right sides. When you drag the drawing objects around on the left you'll see them on the right side. Previously you'd only see the change on the left when you stopped dragging.
|
2025-04-01T06:38:15.823917
| 2016-04-26T19:30:51
|
151219721
|
{
"authors": [
"cunnie",
"ernado",
"vito"
],
"license": "bsd-2-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4861",
"repo": "concourse/atc",
"url": "https://github.com/concourse/atc/pull/77"
}
|
gharchive/pull-request
|
jobs API endpoint serves SVG badges of latest status
@xoebus did the heavy lifting.
typical URI:
https://ci.blabbertabber.com/api/v1/pipelines/main/jobs/always-succeeds/badge
Build Status
Shield
Implemented
nil (build never run)
yes
StatusAborted
yes
StatusSucceeded
yes
StatusFailed
yes
StatusErrored
yes
Similar to the badges served by Travis CI
SVG created courtesy shields.io
Badge style: flat
Here's a deployment of a Concourse dev release with the badges: https://ci.blabbertabber.com:
And here are the badges (pulled from the live site!):
[fixes #72388008]
Signed-off-by: Brian Cunnie<EMAIL_ADDRESS>
looks great! prioritizing.
I need it so bad, looks great.
Will use it definitely :)
|
2025-04-01T06:38:15.851011
| 2018-03-16T12:20:57
|
305907581
|
{
"authors": [
"Mesuva",
"hemipatu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4862",
"repo": "concrete5-community-store/community_store",
"url": "https://github.com/concrete5-community-store/community_store/issues/349"
}
|
gharchive/issue
|
Payment method using coinpayments.net
Hi there, I am wondering if it would be possible to add https://www.coinpayments.net/ as a payment method onto the store as this is a quickly growing form of payment and it would also open more possibilities with what people are accepting etc.
There shouldn't be any restriction from the community store side of things to allow for integration here, it can handle both client side and server side handling in different ways to accommodate different payment gateways.
It really just comes down to the amount of work that might be required. Stripe for example has both a client side library (for the pop up form) as well as strong server side libraries, so it ends up being a bit easier than others. For this one, I'm not really seeing anything that handles the client side, so stuff like picking what cryptocurrency you want to use, really the final payment interface would all need to be written.
Personally I wouldn't have plans to work on this (I'd need a client to request it), but I'm happy to point someone in the right direction. The existing payment gateways do act as good code examples.
If a "client" did decide they wanted said feature. what's your estimates on pricing for it's creation?
Furthermore would it be something you'd consider if you where to be paid to develop this new payment method?
|
2025-04-01T06:38:15.852960
| 2021-04-26T22:26:31
|
868249323
|
{
"authors": [
"mdius"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4863",
"repo": "concrete5-community-store/community_store",
"url": "https://github.com/concrete5-community-store/community_store/issues/610"
}
|
gharchive/issue
|
EasyPost will not load anymore.
www.mdius.com
Just spins on checkout again.
Was trying to delete my EasyPost USPS shipping and it crashed my shipping.
I cannot give error codes till tomorrow, but it is related to EasyPost has no index number.
You've helped me with this issue before.
It seems to be working now.
Was able to delete the problem and shipping works.
|
2025-04-01T06:38:15.863313
| 2019-12-27T22:23:33
|
543007394
|
{
"authors": [
"JohntheFish",
"aembler",
"fmaruna",
"gary-portlandlabs",
"katalysis"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4864",
"repo": "concrete5/concrete5",
"url": "https://github.com/concrete5/concrete5/issues/8322"
}
|
gharchive/issue
|
Update Dark Panels for greater clarity, fresher design palette.
These dark side panels are really hard on some screens/eyes and are really the only place where we go so low light/contrast. I do think it'd be nice if we had a 'dark mode' - but that should be all encompassing and higher contrast. Lets see some options for updating that which is grey on dark grey to a fresher (and more ADA friendly) palette.
Agree with @mlocati re. search block, content type selector menu should be the main event and search box then only show if necessary.
I like the block menu on single lines allowing for longer block names. Could the button container be extended full width as with the container menu above?
Also the popup menu entry to open the content sidebar is overdue a change from 'Add Block' to 'Add Content' (this ties in with the wording used in the top bar '+' icon tooltip).
Should Page Template and Theme even be on the same panel?
There are times when the grid of block icons is useful, and times when the list with room for names is useful. Can we have it toggle between the 2 views?
I like the colour @katalysis has added. More interesting than grey on grey.
Can the current Alt-click behaviour on the add '+' icon to lock the block panel be made more obvious? May users never realise it is there. There are also inconsistencies of panel behaviour when the block add '+' is locked and switching between Blocks/Stacks/Clipboard.
When filtering the block add panel with a search, can the search be sticky, at least for the current page edit, or perhaps extended to the current user session?
Thinking of @fmaruna's liking for my 'recent buttons' quick pick on the Button Nav demo, can something similar be done for the block add panel? A section at the top of the block panel with a quick pick of the last few block types added for easy selection without needing to scroll down or search.
That would be for block types, unpopulated.
A different shortlist would be for the few most recently added/saved blocks to automatically be at the top of the Clipboard, or even another section for the last N blocks added, so there would be separate panels for 'Clipboard>' and 'Recent Blocks>'
Marking this as completed.
|
2025-04-01T06:38:15.864490
| 2017-03-01T14:29:42
|
211097351
|
{
"authors": [
"mlocati"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4865",
"repo": "concrete5/concrete5",
"url": "https://github.com/concrete5/concrete5/pull/5155"
}
|
gharchive/pull-request
|
Faster nginx rewrite rule
As discussed in Slack with @KorvinSzanto
PS: it requires some testing in case of DIR_REL is not empty (ie concrete5 is installed in a subdirectory)
|
2025-04-01T06:38:15.869835
| 2018-07-16T06:54:33
|
341407410
|
{
"authors": [
"mlocati"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4866",
"repo": "concrete5/concrete5",
"url": "https://github.com/concrete5/concrete5/pull/6903"
}
|
gharchive/pull-request
|
Fix infinite redirection visiting existing dirs when seo.trailing_slash is false
Condition:
using Apache with .htaccess set by concrete5
users visits an existing directory (eg /concrete/images)
Execution:
the browser visits /concrete/images
Apache sees that /concrete/images is not a file and it's not a folder containing index.php/index.html, and it sends the request to index.php but with /concrete/images/ as request path
Here's the mod_rewrite log:strip per-dir prefix: /var/www/concrete/images -> concrete/images
applying pattern '.' to uri 'concrete/images'
RewriteCond: input='/var/www/concrete/images' pattern='!-f' => matched
RewriteCond: input='/var/www/concrete/images/index.html' pattern='!-f' => matched
RewriteCond: input='/var/www/concrete/images/index.php' pattern='!-f' => matched
rewrite 'concrete/images' -> 'index.php'
add per-dir prefix: index.php -> /var/www/index.php
trying to replace prefix /var/www/ with /
strip matching prefix: /var/www/index.php -> index.php
add subst prefix: index.php -> /index.php
internal redirect with /index.php [INTERNAL REDIRECT]
strip per-dir prefix: /var/www/concrete/images/ -> concrete/images/
applying pattern '.' to uri 'concrete/images/'
RewriteCond: input='/var/www/concrete/images/' pattern='!-f' => matched
RewriteCond: input='/var/www/concrete/images//index.html' pattern='!-f' => matched
RewriteCond: input='/var/www/concrete/images//index.php' pattern='!-f' => matched
rewrite 'concrete/images/' -> 'index.php'
add per-dir prefix: index.php -> /var/www/index.php
trying to replace prefix /var/www/ with /
strip matching prefix: /var/www/index.php -> index.php
add subst prefix: index.php -> /index.php
internal redirect with /index.php [INTERNAL REDIRECT]
strip per-dir prefix: /var/www/index.php -> index.php
applying pattern '.' to uri 'index.php'
RewriteCond: input='/var/www/index.php' pattern='!-f' => not-matched
pass through /var/www/index.php
the response factory sees that the request URL does not end with a slash, so it build a redirect response to /concrete/images
go to point 1 until the browsers stops the request because of too many redirects
PS: fix #6785
GitHub is having some issues: I pushed https://github.com/mlocati/concrete5/commit/bb60410b8253e79905fa1629a6349843e72b0d65 to my fix-infinite-redirection-existing-dir branch, but this PR isn't updated.
Ok, now it's here too
|
2025-04-01T06:38:15.874943
| 2022-04-28T16:09:35
|
1218956157
|
{
"authors": [
"magnunor"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4867",
"repo": "conda-forge/atomap-feedstock",
"url": "https://github.com/conda-forge/atomap-feedstock/pull/10"
}
|
gharchive/pull-request
|
Release 0.3.2
Checklist
[x] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[x] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
@conda-forge-admin, please rerender
|
2025-04-01T06:38:15.876079
| 2016-06-21T17:02:38
|
161489647
|
{
"authors": [
"conda-forge-linter",
"ocefpaf"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4868",
"repo": "conda-forge/ckanapi-feedstock",
"url": "https://github.com/conda-forge/ckanapi-feedstock/pull/2"
}
|
gharchive/pull-request
|
Update to 3.6
Using GitHub tarball.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
|
2025-04-01T06:38:15.883793
| 2024-08-07T13:50:53
|
2453543813
|
{
"authors": [
"beckermr",
"h-vetinari",
"isuruf"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4869",
"repo": "conda-forge/conda-forge-repodata-patches-feedstock",
"url": "https://github.com/conda-forge/conda-forge-repodata-patches-feedstock/pull/821"
}
|
gharchive/pull-request
|
Don't patch llvm packages for new versions
Checklist
[ ] Used a static YAML file for the patch if possible (instructions).
[ ] Only wrote code directly into generate_patch_json.py if absolutely necessary.
[ ] Ran pre-commit run -a and ensured all files pass the linting checks.
[ ] Ran python show_diff.py and posted the output as part of the PR.
[ ] Modifications won't affect packages built in the future.
cc @beckermr @h-vetinari
The PR is clear and thank you! Can we xref any known issues about this or why we stopped patching now?
I don't remember why I patched it in the first place. Let's try not patching for new versions.
precommit.ci autofix
Thanks.
Can we xref any known issues about this or why we stopped patching now?
https://github.com/conda-forge/llvmdev-feedstock/pull/142, https://github.com/conda-forge/llvmdev-feedstock/issues/141 are the ones I'm aware of. Looking at the git history, this seems to come from #20, which doesn't have other references AFAICT.
|
2025-04-01T06:38:15.890091
| 2021-12-04T00:06:38
|
1071062026
|
{
"authors": [
"conda-forge-linter",
"ickc"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4870",
"repo": "conda-forge/defopt-feedstock",
"url": "https://github.com/conda-forge/defopt-feedstock/pull/11"
}
|
gharchive/pull-request
|
upgrade to 6.2.0
Checklist
[x] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
Hi! This is the friendly automated conda-forge-linting service.
I wanted to let you know that I linted all conda-recipes in your PR (recipe) and found some lint.
Here's what I've got...
For recipe:
requirements: run: colorama>=0.3.4 must contain a space between the name and the pin, i.e. colorama >=0.3.4
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
This fixes a few dependencies that the bot in #10 didn't get.
|
2025-04-01T06:38:15.898665
| 2024-03-08T10:38:17
|
2175753778
|
{
"authors": [
"h-vetinari",
"hmacdope",
"jakirkham",
"zklaus"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4871",
"repo": "conda-forge/dgl-feedstock",
"url": "https://github.com/conda-forge/dgl-feedstock/pull/36"
}
|
gharchive/pull-request
|
DGL v2.1.0 w/ liburing
Checklist
[ ] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
@conda-forge-admin, please rerender
The broken link check is most likely due to conda-forge/admin-requests#958. @conda-forge-admin, please restart ci.
@conda-forge-admin, please rerender
@conda-forge-admin , please restart CI
Ok, seems everything is working now, except for the Osx builds, which are blocked by https://github.com/conda-forge/tensorflow-feedstock/pull/372 or https://github.com/conda-forge/tensorflow-feedstock/pull/374.
@conda-forge-admin, please rerender
@h-vetinari, would you have any hints on what to do about those stdlib2.12 cuda builds?
So I think what's happening is that - in a CUDA-enabled recipe - you're overriding this big zip, and as you're only overriding it partially (one key out of many), you need to match the length of the other keys in that zip.
I needed to do something similar in arrow to override the compiler version (note, you'd need a third entry for linux because the CUDA 12 migrator adds to the length of that zip).
Perhaps an easier alternative would be to set
os_version:
linux_64: cos7
in conda-forge.yml though, which should also ensure that you get the newer stdlib
@conda-forge-admin, please rerender
Seems this is finally ready, @conda-forge/dgl :tada: However, I suggest to first merge #38, then rebase this on top of the new main and only merge after that.
Thanks so much for fixing conflicts @zklaus
|
2025-04-01T06:38:15.902732
| 2017-04-22T01:43:49
|
223532171
|
{
"authors": [
"conda-forge-admin",
"conda-forge-linter"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4872",
"repo": "conda-forge/frosted-feedstock",
"url": "https://github.com/conda-forge/frosted-feedstock/pull/3"
}
|
gharchive/pull-request
|
MNT: Re-render the feedstock [ci skip]
Hi! This is the friendly conda-forge-admin automated user.
I've re-rendered this feedstock with the latest version of conda-smithy (2.3.0) and noticed some changes.
If the changes look good, then please go ahead and merge this PR.
If you have any questions about the changes though, please feel free to ping the 'conda-forge/core' team (using the @ notation in a comment).
Remember, for any changes to the recipe you would normally need to increment the version or the build number of the package.
Since this is an infrastructural change, we don't actually need/want a new version to be uploaded to anaconda.org/conda-forge, so the version and build/number are left unchanged and the CI has been skipped.
Thanks!
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
|
2025-04-01T06:38:15.922707
| 2018-10-05T20:19:06
|
367354741
|
{
"authors": [
"auneri",
"conda-forge-linter"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4874",
"repo": "conda-forge/ipyslurm-feedstock",
"url": "https://github.com/conda-forge/ipyslurm-feedstock/pull/2"
}
|
gharchive/pull-request
|
Release v1.1.0
Checklist
[x] Used a fork of the feedstock to propose changes
[x] Bumped the build number (if the version is unchanged)
[x] Reset the build number to 0 (if the version changed)
[x] Re-rendered with the latest conda-smithy
[x] Ensured the license file is being packaged.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
Closing in favor of #1.
Hi! This is the friendly automated conda-forge-webservice.
It appears you are making a pull request from a branch in your feedstock and not a fork. This procedure will generate a separate build for each push to the branch and is thus not allowed. See our documentation for more details.
Please close this pull request and remake it from a fork of this feedstock.
Have a great day!
Hi! This is the friendly automated conda-forge-webservice.
It appears you are making a pull request from a branch in your feedstock and not a fork. This procedure will generate a separate build for each push to the branch and is thus not allowed. See our documentation for more details.
Please close this pull request and remake it from a fork of this feedstock.
Have a great day!
|
2025-04-01T06:38:15.925916
| 2024-07-31T15:20:29
|
2440320422
|
{
"authors": [
"stephenworsley"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4875",
"repo": "conda-forge/iris-esmf-regrid-feedstock",
"url": "https://github.com/conda-forge/iris-esmf-regrid-feedstock/pull/16"
}
|
gharchive/pull-request
|
Update to v0.11.0
Checklist
[x] Used a personal fork of the feedstock to propose changes
[x] Bumped the build number (if the version is unchanged)
[x] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
@conda-forge-admin, please rerender
|
2025-04-01T06:38:15.935396
| 2021-09-26T21:22:50
|
1007507268
|
{
"authors": [
"conda-forge-linter",
"ivan-gomes"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4876",
"repo": "conda-forge/jupyter-sysml-kernel-feedstock",
"url": "https://github.com/conda-forge/jupyter-sysml-kernel-feedstock/pull/9"
}
|
gharchive/pull-request
|
Update jupyter-sysml-kernel to 0.15.0
Checklist
[X] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[X] Reset the build number to 0 (if the version changed)
[X] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[X] Ensured the license file is being packaged.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
|
2025-04-01T06:38:15.939398
| 2023-10-29T13:06:36
|
1966983668
|
{
"authors": [
"jGaboardi"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4877",
"repo": "conda-forge/legendgram-feedstock",
"url": "https://github.com/conda-forge/legendgram-feedstock/pull/6"
}
|
gharchive/pull-request
|
update recipe for re-render
Checklist
[x] Used a personal fork of the feedstock to propose changes
[x] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[x] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[x] Ensured the license file is being packaged.
This PR:
re-renders the recipe
replace summary, which appears to be for simplejson (?)
@conda-forge-admin, please rerender
@conda-forge-admin, please rerender
|
2025-04-01T06:38:15.943871
| 2023-12-04T16:14:00
|
2024224351
|
{
"authors": [
"adibbley"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4878",
"repo": "conda-forge/libcusolver-feedstock",
"url": "https://github.com/conda-forge/libcusolver-feedstock/pull/6"
}
|
gharchive/pull-request
|
Fix RPATH for libraries.
Checklist
[ ] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
This PR fixes an issue with RPATH entries in libraries.
See conda-forge/cuda-feedstock#10
Requires:
https://github.com/conda-forge/libcublas-feedstock/pull/14
https://github.com/conda-forge/cuda-nvrtc-feedstock/pull/3
https://github.com/conda-forge/libnvjitlink-feedstock/pull/3
|
2025-04-01T06:38:15.951532
| 2024-12-16T14:33:19
|
2742554557
|
{
"authors": [
"SophieCurinier",
"conda-forge-admin"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4879",
"repo": "conda-forge/mira-simpeg-feedstock",
"url": "https://github.com/conda-forge/mira-simpeg-feedstock/pull/1"
}
|
gharchive/pull-request
|
[DEVOPS-601] Update conda recipe with Jinja template
Checklist
[x] Used a personal fork of the feedstock to propose changes
[x] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
@conda-forge-admin, please rerender@conda-forge-admin, please rerender@conda-forge-admin, please rerender@conda-forge-admin, please rerender
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe/meta.yaml) and found it was in an excellent condition.
I do have some suggestions for making it better though...
For recipe/meta.yaml:
ℹ️ noarch: python recipes should usually follow the syntax in our documentation for specifying the Python version.
For the host section of the recipe, you should usually use python {{ python_min }} for the python entry.
For the run section of the recipe, you should usually use python >={{ python_min }} for the python entry.
If the package requires a newer Python version than the currently supported minimum version on conda-forge, you can override the python_min variable by adding a Jinja2 set statement at the top of your recipe (or using an equivalent context variable for v1 recipes).
This message was generated by GitHub Actions workflow run https://github.com/conda-forge/conda-forge-webservices/actions/runs/12354822754. Examine the logs at this URL for more detail.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe/meta.yaml) and found it was in an excellent condition.
|
2025-04-01T06:38:15.969639
| 2023-11-23T17:55:55
|
2008637676
|
{
"authors": [
"johntruckenbrodt"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4882",
"repo": "conda-forge/pyrosar-feedstock",
"url": "https://github.com/conda-forge/pyrosar-feedstock/pull/28"
}
|
gharchive/pull-request
|
version 0.23.0
Checklist
[x] Used a personal fork of the feedstock to propose changes
[x] Bumped the build number (if the version is unchanged)
[x] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[x] Ensured the license file is being packaged.
@conda-forge-admin, please rerender
|
2025-04-01T06:38:15.975081
| 2021-04-25T11:36:28
|
866988583
|
{
"authors": [
"conda-forge-linter",
"jan-janssen",
"ltalirz"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4883",
"repo": "conda-forge/qe-feedstock",
"url": "https://github.com/conda-forge/qe-feedstock/pull/15"
}
|
gharchive/pull-request
|
try running test suite
fix #14
Checklist
[ ] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
@jan-janssen now running test suite
it's only 17 tests, and not all executables are tested but I guess it's a good start
@conda-forge-admin, please rerender
@ltalirz thanks a lot
@ltalirz By the way if you want to add yourself as a maintainer feel free to do so.
|
2025-04-01T06:38:15.979820
| 2022-10-11T03:44:19
|
1403954808
|
{
"authors": [
"conda-forge-linter",
"isuruf",
"mbargull"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4884",
"repo": "conda-forge/r-base-feedstock",
"url": "https://github.com/conda-forge/r-base-feedstock/pull/224"
}
|
gharchive/pull-request
|
[do not merge; testing stuff] 4.1: Reenable Windows builds
Checklist
[ ] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=582673&view=logs&j=a70f640f-cc53-5cd3-6cdc-236a1aa90802 built successfully!
I'm re-enabling the other platforms again.
Thanks
|
2025-04-01T06:38:15.996825
| 2024-10-24T02:06:31
|
2610214698
|
{
"authors": [
"conda-forge-admin",
"sfc-gh-srudenko"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4886",
"repo": "conda-forge/snowflake-ml-python-feedstock",
"url": "https://github.com/conda-forge/snowflake-ml-python-feedstock/pull/11"
}
|
gharchive/pull-request
|
Dependency update
Update dependencies
bump version
Checklist
[ ] Used a personal fork of the feedstock to propose changes
[ ] Bumped the build number (if the version is unchanged)
[ ] Reset the build number to 0 (if the version changed)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[ ] Ensured the license file is being packaged.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe/meta.yaml) and found it was in an excellent condition.
I do have some suggestions for making it better though...
For recipe/meta.yaml:
No valid build backend found for Python recipe for package snowflake-ml-python using pip. Python recipes using pip need to explicitly specify a build backend in the host section. If your recipe has built with only pip in the host section in the past, you likely should add setuptools to the host section of your recipe.
|
2025-04-01T06:38:16.082785
| 2022-09-12T21:50:47
|
1370558945
|
{
"authors": [
"conda-forge-linter",
"tomvothecoder"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4887",
"repo": "conda-forge/xcdat-feedstock",
"url": "https://github.com/conda-forge/xcdat-feedstock/pull/12"
}
|
gharchive/pull-request
|
Add python-dateutil as a dependency
Checklist
[x] Used a personal fork of the feedstock to propose changes
[x] Bumped the build number (if the version is unchanged)
[ ] Re-rendered with the latest conda-smithy (Use the phrase @conda-forge-admin, please rerender in a comment in this PR for automated rerendering)
[x] Ensured the license file is being packaged.
Hi! This is the friendly automated conda-forge-linting service.
I just wanted to let you know that I linted all conda-recipes in your PR (recipe) and found it was in an excellent condition.
@conda-forge-admin, please rerender
|
2025-04-01T06:38:16.084827
| 2020-05-19T07:58:12
|
620762679
|
{
"authors": [
"marcelotrevisani",
"woutdenolf"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4888",
"repo": "conda-incubator/grayskull",
"url": "https://github.com/conda-incubator/grayskull/issues/129"
}
|
gharchive/issue
|
disable compilers when dependencies are extra
For the following package tornado_sqlalchemy_login there is a dependency for pybind11 but it is just an extra and in dev mode. However, grayskull is adding the compilers anyway. It should not add it
grayskull is fetching the extra's but not using them yet (see https://github.com/conda-incubator/grayskull/issues/150)
So how is grayskull "adding" the compilers and in what section?
that is correct now, and thanks again for the feature that you developed :)
|
2025-04-01T06:38:16.089810
| 2023-09-25T20:22:30
|
1912226039
|
{
"authors": [
"kenodegard"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4889",
"repo": "conda/actions",
"url": "https://github.com/conda/actions/pull/127"
}
|
gharchive/pull-request
|
Add template-files composite action
Description
We've been using github.com/BetaHuhn/repo-file-sync-action to synchronize files from conda/infrastructure for a while now and have found that it doesn't perform in the way we need it to (bulk syncing with templating), the upstream author is not particularly responsive anymore, and since the upstream action is written in NodeJS/Typescript its not particularly simple to fork and fix.
Thus we implement a small composite action that implements a pull process and uses Jinja as the templating engine. Since this is a decentralized pull process this also gives more control to individual repos instead of having a central sync conductor.
I have verified that the workflow still works in https://github.com/conda-sandbox:
https://github.com/conda-sandbox/upstream mocks https://github.com/conda/infrastructure
https://github.com/conda-sandbox/downstream mocks any of the conda projects, e.g.: https://github.com/conda/conda
See a scheduled run (where new updates are fetched from upstream): https://github.com/conda-sandbox/downstream/pull/19
And a manually triggered run (where we enable nested templating features): https://github.com/conda-sandbox/downstream/pull/20
|
2025-04-01T06:38:16.120355
| 2022-11-22T20:18:06
|
1460497674
|
{
"authors": [
"maksenius",
"voscob"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4890",
"repo": "conduitio-labs/conduit-connector-stripe",
"url": "https://github.com/conduitio-labs/conduit-connector-stripe/issues/7"
}
|
gharchive/issue
|
Feature: Source connector modes
Feature description
Add ability source connector to working in following modes:
read existing data, enable CDC
ignore existing data, enable CDC
It was added in this PR:
https://github.com/conduitio-labs/conduit-connector-stripe/pull/8
|
2025-04-01T06:38:16.133879
| 2020-04-14T14:47:07
|
599630036
|
{
"authors": [
"PaluruSumanth",
"mhowlett"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4891",
"repo": "confluentinc/confluent-kafka-dotnet",
"url": "https://github.com/confluentinc/confluent-kafka-dotnet/issues/1248"
}
|
gharchive/issue
|
Apache.Avro support for .net framework 4.x
Why is that .Net framework support is removed in Apache.Avro 1.9.1 onwards? Is there plan to support .Net framework in upcoming versions? Is there anyway for .Net framework 4.x applications to consume latest apache.avro changes?
previously, Confluent maintained a fork of the official repo to facilitate rapidly making changes required to use the library with Confluent.Kafka. In 1.9.1, the official Avro library incorporated all our bug fixes and overtook our fork in terms of features and we switched over in 1.4. Our fork supported net451, but the official library only supports back to netstandard2.0. This is compatible with net461 and above.
|
2025-04-01T06:38:16.140338
| 2020-07-10T09:17:08
|
654641695
|
{
"authors": [
"gelihu",
"mhowlett"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4892",
"repo": "confluentinc/confluent-kafka-dotnet",
"url": "https://github.com/confluentinc/confluent-kafka-dotnet/issues/1350"
}
|
gharchive/issue
|
Q: transactional producer configs for mulit-broker(cluster) behind load balancer
Description
I am trying following:
Setup Two brokers behind a load balancer(which will route traffic based on availability, load, etc).
Create a producer(with transaction support) by setting load balancer host address as bootstrap server.
It works fine fine till there are no disconnects, but one fine day following happens:
%6|1591244167.286|FAIL|app-trx#producer-2| [thrd:sasl_plaintext://load-balanacer-host:9092/bootstrap]: sasl_plaintext://load-balanacer-host:9092/bootstrap: Disconnected (after 1516513ms in state UP)
%3|1591244167.286|ERROR|app-trx#producer-2| [thrd:sasl_plaintext://load-balanacer-host:9092/bootstrap]: sasl_plaintext://load-balanacer-host:9092/bootstrap: Disconnected (after 1516513ms in state UP)
%3|1591244453.733|FAIL|app-trx#producer-2| [thrd:sasl_plaintext://broker-1-ip:9092/1001]: sasl_plaintext://broker-1-ip:9092/1001: Connect to ipv4#broker-1-ip:9092 failed: No route to host (after 3006ms in state CONNECT)
%3|1591244453.733|ERROR|app-trx#producer-2| [thrd:sasl_plaintext://broker-1-ip:9092/1001]: sasl_plaintext://broker-1-ip:9092/1001: Connect to ipv4#broker-1-ip:9092 failed: No route to host (after 3006ms in state CONNECT)
%3|1594301839.904|ADDPARTS|app-trx#producer-2| [thrd:main]: TxnCoordinator/1002: Failed to add partition "topic-name" [0] to transaction: Broker: Producer attempted to use a producer id which is not currently assigned to its transactional id
%1|1594301839.904|TXNERR|app-trx#producer-2| [thrd:main]: Fatal transaction error: Failed to add partitions to transaction: Broker: Producer attempted to use a producer id which is not currently assigned to its transactional id (INVALID_PRODUCER_ID_MAPPING)
%3|1594301839.904|ERROR|app-trx#producer-2| [thrd:main]: Fatal error: Broker: Producer attempted to use a producer id which is not currently assigned to its transactional id: Failed to add partitions to transaction: Broker: Producer attempted to use a producer id which is not currently assigned to its transactional id
Did I configure something wrong, please advice?
I did get on zkCli for /brokers/topics/topic-name, it gave following:
{"version":1,"partitions":{"1":[1001,1002],"0":[1002,1001]}}
How to reproduce
Not an issue.
Checklist
Please provide the following information:
[ ] A complete (i.e. we can run it), minimal program demonstrating the problem. No need to supply a project file.
[x] Confluent.Kafka nuget version. 1.4.0
[ ] Apache Kafka version.
[ ] Client configuration.
[ ] Operating system.
[ ] Provide logs (with "debug" : "..." as necessary in configuration).
[ ] Provide broker log excerpts.
[ ] Critical issue.
Setup Two brokers behind a load balancer(which will route traffic based on availability, load, etc).
Kafka + protocol take care of availability / load balancing for you already - you shouldn't try to use a load balancer for this purpose. The .net client expects to maintain connections directly to brokers as it requires. I'm aware of some people using a level of indirection on the bootstrap servers to allow them to switch out broker addresses (if all other connections fail, the bootstrap server addresses will be used to find the cluster again), but I haven't done this personally, and I believe there is some devil in the detail here.
|
2025-04-01T06:38:16.146352
| 2020-11-04T17:24:29
|
736292562
|
{
"authors": [
"AdikaSinghvi",
"mhowlett"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4893",
"repo": "confluentinc/confluent-kafka-dotnet",
"url": "https://github.com/confluentinc/confluent-kafka-dotnet/issues/1448"
}
|
gharchive/issue
|
Not able to send messages to kafka asynchronously and get a callback to handle the errors (if any) later
Description
We are using Confluent.Kafka 1.5.2. For serializing keys and values, we use Confluent.SchemaRegistry.Serdes.Avro 1.5.2.
Apache.Avro contract version is 1.9.1.
We are trying to use the below Produce API to send messages asynchronously and get a callback at a later point. We plan to handle the errors via callback.
void Produce(string topic, Message<TKey, TValue> message, Action<DeliveryReport<TKey, TValue>> deliveryHandler = null);
At runtime, we are getting error "System.InvalidOperationException: Produce called with an IAsyncSerializer key serializer configured but an ISerializer is required.
at Confluent.Kafka.Producer2.Produce(TopicPartition topicPartition, Message2 message, Action`1 deliveryHandler)"
How to reproduce
producerBuilder = new ProducerBuilder<string, T>(producerconfig)
.SetKeySerializer(new Confluent.SchemaRegistry.Serdes.AvroSerializer(schemaRegistryconfig))
.SetValueSerializer(new Confluent.SchemaRegistry.Serdes.AvroSerializer(schemaRegistryconfig))
.Build();
producerBuilder.Produce(topicName, data, CallbackHandler);
Checklist
Please provide the following information:
[ ] A complete (i.e. we can run it), minimal program demonstrating the problem. No need to supply a project file.
[x] Confluent.Kafka nuget version. 1.5.2
[ ] Apache Kafka version.
[ ] Client configuration.
[ ] Operating system.
[ ] Provide logs (with "debug" : "..." as necessary in configuration).
[ ] Provide broker log excerpts.
[x] Critical issue.
you need to convert the async serializer into a sync one. there's a helper method for this:
using Confluent.Kafka.SyncOverAsync;
then use .AsSyncOverAsync() on the serializer instance.
|
2025-04-01T06:38:16.188814
| 2022-02-11T09:23:41
|
1132170619
|
{
"authors": [
"buznyusz",
"nsharma-git"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4894",
"repo": "confluentinc/cp-ansible",
"url": "https://github.com/confluentinc/cp-ansible/issues/912"
}
|
gharchive/issue
|
cp-ansible ignores ssl.principal.rules ( w/ MTLS AUTH ) just extract the full subject from the JKS/certificate to super.users
hello,
we want to use MTLS auth between kafka brokers, and trying to use principal.rules to extract data from the certificates subject, but the CP-Ansible ignores this paramater and export the full subject and adding it to the super.users list ( with manual adding it's OK )
https://github.com/confluentinc/cp-ansible/blob/fd56742ff5b63a75d4be64b7c5d4118b68fbe2ee/roles/kafka_broker/tasks/set_principal.yml#L22-L55
here is the lack of this logic !
pls fix it !
Thanks
Hello @buznyusz
Can you please have a look at the PR https://github.com/confluentinc/cp-ansible/pull/905
This is available in 7.1.x onwards.
Let me know if this serves the purpose here. Thanks!
|
2025-04-01T06:38:16.219931
| 2022-02-15T13:23:15
|
1138689560
|
{
"authors": [
"astubbs"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4895",
"repo": "confluentinc/parallel-consumer",
"url": "https://github.com/confluentinc/parallel-consumer/pull/197"
}
|
gharchive/pull-request
|
feature: Retry count available in header
[x] docs
[x] update / link to from issues
[x] include issue number
[x] Remove WorkContainer from any public API - e.g. delay provider in options
Spirit of:
#65
Redundant by
#216
Blocked by:
#223
Redundant by
#216
|
2025-04-01T06:38:16.236165
| 2016-06-03T17:20:18
|
158414165
|
{
"authors": [
"coniel",
"scheung38"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4896",
"repo": "coniel/meteor-react-form-handler",
"url": "https://github.com/coniel/meteor-react-form-handler/issues/14"
}
|
gharchive/issue
|
Login only form
How to modify such that there is only Login, but no Registration form because it is an internal app I am working on and the logic is simply when username and password (hashed) match initialized username and password that is created once upon app launch. Thanks
The form would be basically the same just with fewer fields:
var loginFormSchema = new SimpleSchema({
username: {
type: String,
max: 60,
label: "Username"
},
password: {
type: String,
max: 60,
min: 8,
label: "Password"
}
});
RegistrationForm = React.createClass({
_onSubmit(doc) {
Meteor.loginWithPassword(doc.username, doc.password, (error, result) => {
// Handle error/success
})
},
render() {
return (
<Form schema={registrationFormSchema} id="login-form" onSubmit={this._onSubmit}>
<TextInput name="username" />
<TextInput name="password" type="password" />
<SubmitButton label="Login" />
</Form>
)
}
});
Near the bottom, do you mean
Thanks
|
2025-04-01T06:38:16.244714
| 2024-02-08T01:00:11
|
2124174781
|
{
"authors": [
"acbeni",
"paul-sachs"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4897",
"repo": "connectrpc/connect-query-es",
"url": "https://github.com/connectrpc/connect-query-es/issues/336"
}
|
gharchive/issue
|
enabled param is not taken into consideration with infinitqueries
Hi, There was an issue about enabled param not working in useQuery that seems to have been fixed in this commit, unfortinatenly the problem still persist at useInfinitQuery and also headers provided to useInfinitQuery dont seem to be sent, but maybe the issues are related.
Can yo check this issue please
Yeah looks like I forgot to port that to useInfiniteQuery. Will submit a fix shortly. In the meantime, you can use the 'disableQuery' symbol as the input to that api.
Regarding headers not being passed through, how are you passing the headers in?
As for the headers, i think it was just due to the disable param not being considered. Since the query is supposed to be enabled only after some auth headers are computed. Thats why i didnt find them in the logs. So the headers are fine i think. Its just the disabled param that needs fixing 👍
Thank you @paul-sachs for the fast action 🎉🎉🎉 , any rough estimate on where this will be released ?
Should be able to get to a release later today. Just trying to nail down if a previous change warrants a 2.0 since it changed the queryKey used by infiniteQuery.
Awesome 😎, thank you very much @paul-sachs 🙏
Fixed in https://github.com/connectrpc/connect-query-es/releases/tag/v1.2.0
|
2025-04-01T06:38:16.251198
| 2022-08-23T18:08:49
|
1348349771
|
{
"authors": [
"jakekidd",
"sanchaymittal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4898",
"repo": "connext/nxtp",
"url": "https://github.com/connext/nxtp/issues/1761"
}
|
gharchive/issue
|
As a user, I need my message to be processed on its destination domain
Problem
See https://github.com/connext/nxtp/issues/1714
Subtask: proveAndProcess
We need to implement relayer transmission cycle for proveAndProcess
Ideas to solve this
Implementation plan:
[subgraph] track dispatch() messages at Origin domain subgraph
[subgraph] track proveAndProcess() messages at Destination domain subgraph
[cartographer] update and add new fields to Cartographer
[lighthouse] In the lighthouse, pull all the transfers from cartographers where processed: false | null, and call proveAndProcess() it
Once this will start working we will create a new user story from the dev perspective to add sanity checks for root submission, or additional checks depending on domains and have an ideal time around calling proveAndProcess() and polling.
Subtasks
[x] https://github.com/connext/nxtp/issues/1753
[x] https://github.com/connext/nxtp/issues/1791
[x] https://github.com/connext/nxtp/issues/1792
[x] https://github.com/connext/nxtp/issues/1793
Acceptance Criteria
slow path transfers are working.
completed
|
2025-04-01T06:38:16.261953
| 2019-08-24T05:31:18
|
484781199
|
{
"authors": [
"conorhastings",
"jamesplease"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4899",
"repo": "conorhastings/use-reducer-with-side-effects",
"url": "https://github.com/conorhastings/use-reducer-with-side-effects/issues/13"
}
|
gharchive/issue
|
Feature request: composeReducers
Type
Feature request
Wot's the request
I think it'd be pretty cool if this library introduced a new method, composeReducers, that managed composing multiple reducers into one reducer.
More info
With a typical reducer API, like the one used in useReducer, composing reducers is straightforward. One way to do it is:
export default function composeReducers(reducers) {
return (state, action) =>
reducers.reduceRight(
(prevState, reducer) => reducer(prevState, action),
state
);
}
It is quite a bit more complex to compose reducers with use-reducer-with-side-effects. A few of the things you have to keep in mind are:
the return value of a reducer is on longer the input for the next one, since they receive state as an input but output both state and side effects
since the sideEffects don't pass through to the next reducer, you must keep and concatenate an array of all of the returned side effects outside of the loop
a Symbol is used for NoUpdates in this lib, rather than the previous state object, so you must track the # of Symbols returned to determine if an update actually occurred
None of these points are insurmountable, but it is a bit to think about as a consumer of this lib who is trying to split out a big reducer into smaller pieces. A composeReducers export from this lib would make it easier for devs to organize their reducers in this way.
Other notes
If this isn't added, then it might be a good idea to explicitly export NO_UPDATE_SYMBOL so that folks can reliably check for no updates when writing their own composer fn. (It can be accessed by calling NoUpdate() but that seems a lil hacky for this purpose imo)
Prior art
Redux - composeReducers
Here's an example implementation in userland that seems to be working. It could probably be tidied up a bit with direct access to NO_UPDATE_SYMBOL.
If you're :+1: to this idea, I can make a PR. Lmk what you think!
love the idea, have been thinking about implementation myself
i published 2.0.0 on the beta tag witth . all changes integrateed
Closing since this is resolved :v:
|
2025-04-01T06:38:16.282346
| 2022-03-10T10:36:08
|
1165045218
|
{
"authors": [
"kaiko1203",
"makoshan",
"samisbakedham",
"zengbing15"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4900",
"repo": "consenlabs/token-profile",
"url": "https://github.com/consenlabs/token-profile/pull/7328"
}
|
gharchive/pull-request
|
Add CNDL
Please check the following to ensure a successful verification of Token Profile
.json and .png document checklists
[x] logo dimensions: 120*120 px
[x] logo background: Transparent
[x] At least one transaction on the token contract
[x] For ERC-20 project teams: Upper and lower capitalized alphabets in token address (For checksum purposes)
[x] Same document name for both .json and .png (Including capitalization)
Project background:
Official announcement: https://twitter.com/candle_labs
Project team background: https://candlelabs.org
Project basic information: Decentralized Human-focused metaverse - Candle (CNDL) is the utility and governance token for the Candle Protocol. This protocol is designed to decentralize social media and build out a community governed and moderated place to discuss and share thoughts and ideas.
Media publications: CoinTelegraph
Tradeable exchanges: Decoin, Tokpie, BankCEX, Coinbase
Official Announcement:
Twitter Link
Recommended Gas Limit (For ERC-20 project teams) for transaction: 40,000
1.Your PR is missing project information
2.Official Twitter is missing imToken listing announcement
3.Please don't upload more than one logo
1.Your PR is missing project information
2.Official Twitter is missing imToken listing announcement
3.Please don't upload more than one logo
Fixing now
Twitter Link
Please merge
Merge me
Please merge
Please do not delete in the local database
Fixed it
Please do not delete in the local database
Good to merge? Thanks
You have not uploaded the LOGO, please do not make changes in your database
You have not uploaded the LOGO, please do not make changes in your database
Fixed
Not sure if the project is still being maintained。
提交信息有问题,我先关闭这个 PR
|
2025-04-01T06:38:16.334613
| 2022-11-01T12:19:22
|
1431338679
|
{
"authors": [
"viceice"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4903",
"repo": "containerbase/base",
"url": "https://github.com/containerbase/base/pull/575"
}
|
gharchive/pull-request
|
feat!: rename user to ubuntu
Rename default user from user back to ubuntu
closes #572
Surprised to learn we have some hardcoded PATHs though
yes, there the env isn't yet available 🙃
|
2025-04-01T06:38:16.350638
| 2021-03-12T21:50:47
|
830523277
|
{
"authors": [
"claudiubelu",
"jayunit100",
"jsturtevant"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4904",
"repo": "containerd/containerd",
"url": "https://github.com/containerd/containerd/issues/5188"
}
|
gharchive/issue
|
Windows Containers are leaked when container fails to start
Description
On Windows, when passing a command to a container to run, if the command isn't valid then containerd gives error and clean up its resources but HCS still shows a container is on the system. If you re-run container with the same name you get an error.
Steps to reproduce the issue:
ctr image pull k8s.gcr.io/pause:3.4.1
ctr.exe run --rm k8s.gcr.io/pause:3.4.1 test-1 echo c:\License.txt
ctr: hcsshim::System::CreateProcess test-1: The system cannot find the file specified.
(extra info: {"CommandLine":"echo c:\\License.txt","User":"ContainerUser","Environment":{"PATH":"C:\\Windows\\system32;C:\\Windows;"},"CreateStdInPipe":true,"CreateStdOutPipe":true,"CreateStdErrPipe":true}): unknown
Containerd doesn't know about the container:
ctr c ls
CONTAINER IMAGE RUNTIME
ctr t ls
TASK PID STATUS
but hcs has the container in the list:
hcsdiag list
test-1
Windows Server Container, Unknown
Try to run a valid command:
ctr.exe run --rm k8s.gcr.io/pause:3.4.1 test-1 cmd /c echo c:\License.txt
ctr: hcsshim::CreateComputeSystem test-1: A virtual machine or container with the specified identifier already exists.
(extra info: {"Owner":"containerd-shim-runhcs-v1.exe","SchemaVersion":{"Major":2,"Minor":1},"Container":{"Storage":{"Layers":[{"Id":"e67d9508-fe76-54e0-b098-cabfb7356fd7","Path":"C:\\ProgramData\\containerd\\root\\io.containerd.snapshotter.v1.windows\\snapshots\\24"},{"Id":"e
3732ab4-a8b2-5962-8e88-a0a76ee83d86","Path":"C:\\ProgramData\\containerd\\root\\io.containerd.snapshotter.v1.windows\\snapshots\\23"},{"Id":"51d377df-069c-5b3a-8cb0-e86be7e7cfbe","Path":"C:\\ProgramData\\containerd\\root\\io.containerd.snapshotter.v1.windows\\snapshots\\22"},
{"Id":"994b900c-05b9-5a61-b236-35342d301241","Path":"C:\\ProgramData\\containerd\\root\\io.containerd.snapshotter.v1.windows\\snapshots\\21"}],"Path":"\\\\?\\Volume{0cdda4ac-837b-11eb-b279-845c959994a0}\\"},"Networking":{"AllowUnqualifiedDnsQuery":true,"Namespace":"45B223AA-1
9BF-40D6-9F54-9578BA121B94"}},"ShouldTerminateOnLastHandleClosed":true}): unknown
Describe the results you received:
Describe the results you expected:
What version of containerd are you using:
Tried with 1.4 and 1.5
$ containerd --version
containerd github.com/containerd/containerd v1.4.3 269548fa27e0089a8b8278fc4fc781d7f65a939b
containerd github.com/containerd/containerd v1.5.0-beta.3 02334356d0774a5b194e67b5f1383fd2485ea67a
Any other relevant information (runC version, CRI configuration, OS/Kernel version, etc.):
runc --version
$ runc --version
crictl info
$ crictl info
uname -a
$ uname -a
cmd /c ver
Microsoft Windows [Version 10.0.17763.1577
/cc @kevpar
@claudiubelu
I've tried this as well with containerd v1.4.3, and can confirm the issue. I've also tried something similar with crictl, but that's not an issue there, the failed container still shows up when listing the containers (crictl --runtime-endpoint=npipe://./pipe/containerd-containerd ps), can be removed, and another can be created in its stead, which can then be started.
I've tried it out using the latest main version and latest hcsshim, and this problem doesn't occur anymore:
PS C:\tmp> ctr.exe --address //./pipe//run/containerd-test/containerd run --rm k8s.gcr.io/pause:3.4.1 test-1 echo c:\License.txt
ctr: hcs::System::CreateProcess test-1: The system cannot find the file specified.: unknown
PS C:\tmp> ctr.exe --address //./pipe//run/containerd-test/containerd run --rm k8s.gcr.io/pause:3.4.1 test-1 cmd /c echo c:\License.txt
c:\License.txt
PS C:\tmp> ctr.exe --address //./pipe//run/containerd-test/containerd container list
CONTAINER IMAGE RUNTIME
I've added a test that will make sure we don't have any regressions on this: https://github.com/containerd/containerd/pull/5578
I think we can close this issue. As seen in the previous comments, we haven't seen this anymore, and we have a test for it as well.
/close
|
2025-04-01T06:38:16.368310
| 2017-03-16T20:06:19
|
214823103
|
{
"authors": [
"AkihiroSuda",
"crosbymichael",
"dmcgowan",
"estesp",
"kunalkushwaha",
"samuelkarp",
"sipsma",
"stevvooe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4905",
"repo": "containerd/containerd",
"url": "https://github.com/containerd/containerd/issues/633"
}
|
gharchive/issue
|
Proposal: Add Clone function on Snapshotter
Proposal
Add a Clone function to the Snapshotter interface which can capture the state of an active snapshot. The output is always a writable active snapshot. It can be immediately committed without making any changes or may further altered and commit. Like any other active snapshot, it must be committed or removed. Clone on a committed snapshot allows renaming of committed snapshot without modification.
Additionally formalize that Commit finalizes an active snapshot, not allowing Remove after Commit or multiple calls to Commit on the same snapshot. This is already the current behavior of the overlay driver and should be formalized in the interface. With the existence of the Snapshot function there is no need to support multiple calls to Commit. The use case for multiple Commit would be covered by calling Snapshot + Commit.
Diff
// After commit, the snapshot identified by key is removed.
Commit(ctx context.Context, name, key string, opts ...Opt) error
+ // Clone copies a snapshot to a new snapshot of the same type.
+ //
+ // An active clone will create a new active snapshot with the same
+ // content but changes in one will not be reflected in the other.
+ // A committed clone will create a new read only snapshot of the
+ // same content.
+ //
+ // Options:
+ // WithParent defines the parent to use for the target.
+ // If the target parent differs from the source, the snapshotter
+ // may either rebase or return ErrNotImplemented
+ Clone(ctx context.Context, target, source string, opts ...Opt) error
+
// Remove the committed or active snapshot by the provided key.
WithParent option
Used to support parent verification or rebasing. If the parent does not exist, then the operation must error with ErrNotFound. If the parent exists, but is not the source parent, the snapshotter must either perform a rebase (applying the changes between source and source's parent to the parent) or return ErrNotImplemented (should not return ErrInvalidArgument).
WithLabels option
The new snapshot will use the labels provided through the WithLabels option. Options should not be copied over from the source
Cross namespace copying
A source namespace could be provided on the context to support cloning across namespaces. In this case the source and target may be the same name. Using the same source and target within a single namespace will always return ErrAlreadyExists.
Snapshot flow diagram changes
I diagrammed what the changes may look like to the snapshot model.
https://docs.google.com/a/docker.com/presentation/d/1LYVEkP6VdYV96IO5lhxTwujtpilhzNPAouO0qGUa4z8/edit?usp=sharing
Overlay implementation
When Clone is called in active overlay snapshot, the content of the snapshotted upper directory will be copied to a new active. When Clone is called on a commited overlay snapshot, the new snapshot will just point to the same on disk layer.
Btrfs implementation
Just uses snapshots how Commit was done previously. As part of this change Commit will also remove the original source after the snapshot is completed. Cloning a committed snapshot will just point to the original source.
google doc seems private :sweat_smile:
@AkihiroSuda updated, didn't realize default scope even with a share link was org restricted. Please give it a try.
Thanks for drawing new diagrams! A few questions:
Would every Snapshotter be required to implement an online, mutable copy as part of Snapshot?
On slide 5, what makes undefined(a, P1) different from Prepare(a, P1)? Is it the fact that P2's parent is P0?
We decided this dual nature was kind of confusing so left it open as a possible future function, name yet to be defined. [...] Should I just remove the slide?
I think that makes sense unless it's something we still want to solve for 1.0.0. In my view, it's better to document the design that we're implementing than future potential options that we're not.
@samuelkarp agreed, when we update the design we will not include that slide. Just made the slides as a place to capture the diagrams for the future doc update.
So does this Snapshot function is just to support Multiple Commit use case?
Also this means consumers of Snapshot drivers, must decide prior if they want to make any changes after commit. e.g. Image builder should use snapshot + commit for intermediate builds and may use Commit to squash the image.
Also, what will be property of Commit layer. i.e. it will be equals to View or View is required to read only operations too?
This has been implemented as a subcommand in ctr
Renaming this, it threw me off as well. This is about a potential function on the Snapshotter interface not ctr snapshot.
Looking at this with experience eyes, I think there is a definite use case for this proposal for backups and other online snapshotting.
I think for sanity's sake, we should consider calling it Clone or Copy.
Adding to the 1.2 milestone.
Looking at this with experience eyes, I think there is a definite use case for this proposal for backups and other online snapshotting.
Do we support using a c8d snapshot as a persistent data volume?
Updated this with more details, considering this for 1.2 to support BuildKit cross namespace exporting
Mostly LGTM on this proposal. Glad its here.
A source namespace could be provided on the context to support cloning across namespaces. In this case the source and target may be the same name. Using the same source and target within a single namespace will always return ErrAlreadyExists.
Might want to detail this workflow a little more. How are the source and target namespaces specified?
This hasn't been in a milestone for a few releases; do we want to resurrect this and possibly see if there is interest in the community to work on it?
Just to chime in with some community interest, I just finished up a bunch of work for the upcoming Buildkit release that required extending the containerd snapshotter interface with a method called Merge. Merge is different than Clone, but has a lot of conceptual overlap with the WithParent rebasing feature described here and probably could have taken advantage of Clone if it had existed. If interested, background on the Buildkit features that required Merge are in these docs.
While I can't just literally upstream Merge and rename the method to Clone due to the differences, it would still be very nice to be able to use Clone in Merge's implementation if nothing else, so if there's still interest in Clone existing I could potentially help out with it (time permitting of course).
|
2025-04-01T06:38:16.371736
| 2024-06-18T23:22:59
|
2360902250
|
{
"authors": [
"cpuguy83",
"thaJeztah"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4906",
"repo": "containerd/containerd",
"url": "https://github.com/containerd/containerd/pull/10360"
}
|
gharchive/pull-request
|
1.7: Add back support for OTLP config from toml
This was broken due to 9360e37169f2ba3135f7a6f39a3ab7c9231abbd6 (backport from main).
The backport should not have broken the toml config so this more or less restores that.
This treats the env vars as preferred.
It is not perfect since, as an example, Endpoint could be set on the config and the env to set the protocol could be set and these could conflict, however I think this is an unlikely scenario and people should use one or the other, not both.
This change converts the config into the relevant environment variables before initializing tracers.
Fixes #10358
cc @vvoland
@cpuguy83 may need a similar patch for 1.6 as well, or at least I see there was a backport to 1.6 as well;
https://github.com/containerd/containerd/pull/9993
|
2025-04-01T06:38:16.379195
| 2018-11-09T22:53:07
|
379349925
|
{
"authors": [
"Random-Liu",
"codecov-io"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4907",
"repo": "containerd/containerd",
"url": "https://github.com/containerd/containerd/pull/2774"
}
|
gharchive/pull-request
|
[release/1.1] cherry-pick: enhance: update v1/v2 runtime
Cherrypick of https://github.com/containerd/containerd/pull/2769.
For https://github.com/containerd/containerd/issues/2709.
Only v1 change is cherry-picked.
Signed-off-by: Wei Fu<EMAIL_ADDRESS>
Codecov Report
Merging #2774 into release/1.1 will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## release/1.1 #2774 +/- ##
============================================
Coverage 48.99% 48.99%
============================================
Files 85 85
Lines 7603 7603
============================================
Hits 3725 3725
Misses 3203 3203
Partials 675 675
Flag
Coverage Δ
#linux
48.99% <ø> (ø)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update b337430...665815b. Read the comment docs.
|
2025-04-01T06:38:16.397546
| 2019-10-02T09:25:17
|
501375953
|
{
"authors": [
"Zyqsempai",
"codecov-io",
"crosbymichael",
"dmcgowan",
"estesp",
"fuweid"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4908",
"repo": "containerd/containerd",
"url": "https://github.com/containerd/containerd/pull/3709"
}
|
gharchive/pull-request
|
Added filters to snapshots API
Fixes #3708
Added filters to snapshots and walk function interface.
@Zyqsempai thanks for working on this. Can you sign your commits with the proper DCO? https://github.com/containerd/project/blob/master/CONTRIBUTING.md#sign-your-work
Each commit will need to be signed, feel free to keep them separate or squash and force push your branch.
The vendored snapshotters (aufs, zfs) need to be updated to the new Walk(..) signature before this change can pass CI:
cmd/containerd/builtins.go:1: vendor/github.com/containerd/aufs/aufs.go:81:9: cannot use &(snapshotter literal) (value of type *snapshotter) as snapshots.Snapshotter value in return statement: wrong type for method Walk (typecheck)
Of course there is a bit of a chicken-and-egg problem in getting those fixed and re-vendored without this change to the signature being merged--might have to temporarily point their vendors to a branch of containerd with the change?
@estesp I opened PR's for them also.
Codecov Report
Merging #3709 into master will increase coverage by 3.4%.
The diff coverage is 50.74%.
@@ Coverage Diff @@
## master #3709 +/- ##
=========================================
+ Coverage 41.98% 45.38% +3.4%
=========================================
Files 131 118 -13
Lines 14536 11699 -2837
=========================================
- Hits 6103 5310 -793
+ Misses 7525 5476 -2049
- Partials 908 913 +5
Flag
Coverage Δ
#linux
45.38% <50.74%> (-0.03%)
:arrow_down:
#windows
?
Impacted Files
Coverage Δ
snapshots/storage/bolt.go
59.12% <10%> (-0.6%)
:arrow_down:
snapshots/btrfs/btrfs.go
57.39% <100%> (-0.9%)
:arrow_down:
snapshots/devmapper/snapshotter.go
66.39% <100%> (ø)
:arrow_up:
snapshots/native/native.go
52.68% <100%> (+10.26%)
:arrow_up:
snapshots/overlay/overlay.go
52.52% <100%> (ø)
:arrow_up:
metadata/snapshot.go
56.44% <50%> (+8.62%)
:arrow_up:
metadata/adaptors.go
54.63% <90.47%> (+14.16%)
:arrow_up:
remotes/docker/fetcher.go
42.5% <0%> (-7.5%)
:arrow_down:
remotes/docker/auth.go
63.82% <0%> (-3.97%)
:arrow_down:
remotes/docker/status.go
21.42% <0%> (-3.58%)
:arrow_down:
... and 83 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 9c86b8f...ee085d7. Read the comment docs.
Basically, LGTM. But the CI green is not trust because the vendor check was broken 6days ago. I think we can remove the out of tree snapshotter plugins from vendor.conf shortly and then add it back after this one merged. is it ok?
@fuweid I agree, I was going to do that and forgot. Commented those out from vendor for now
|
2025-04-01T06:38:16.400837
| 2022-04-27T18:12:51
|
1217716862
|
{
"authors": [
"log1cb0mb",
"thaJeztah"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4909",
"repo": "containerd/containerd",
"url": "https://github.com/containerd/containerd/pull/6865"
}
|
gharchive/pull-request
|
Bump opencontainers/selinux from 1.10.0 to 1.10.1
Full Changelog:
https://github.com/opencontainers/selinux/compare/v1.10.0...v1.10.1
Specifically patch https://github.com/opencontainers/selinux/pull/173 which fixes the issue: https://github.com/containerd/containerd/issues/6767
Signed-off-by: Nabeel Rana<EMAIL_ADDRESS>
Looks like you may have to run go mod tidy from within the integration/client directory as well (which also has a go.mod);
Files /home/runner/work/containerd/containerd/src/github.com/containerd/containerd/integration/client/go.mod and /tmp/tmp.PYcoKntieN/containerd/integration/client/go.mod differ
Files /home/runner/work/containerd/containerd/src/github.com/containerd/containerd/integration/client/go.sum and /tmp/tmp.PYcoKntieN/containerd/integration/client/go.sum differ
/ok-to-test
|
2025-04-01T06:38:16.405230
| 2023-01-03T19:18:16
|
1517815074
|
{
"authors": [
"Zheaoli"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4910",
"repo": "containerd/nerdctl",
"url": "https://github.com/containerd/nerdctl/pull/1797"
}
|
gharchive/pull-request
|
[refactor] Add global flag process helper function
I'm so sorry I have to make a huge PR again.
After #1793, we have a types.GlobalCommandOptions, and for #1792 and #1791 and many refactor PR int he future, we may need to introduce a new helper function to process the global flag.
In this PR, I introduce the new helper function and use this function to process the global flag in all subcommand.
There may be some duplicated code for some subcommand like nerdctl run, I think this should be not a problem, I will merge it when I refactor the run command,
Signed-off-by: Zheao.Li<EMAIL_ADDRESS>
(Not request changes)
In the current PR, most commands are refactored into a half-way state (e.g., global flag is processed (sometimes, repeatedly) whenever it's needed. Ideally, it should only be processed once at the entry of the command, and let other usage use the processed struct.
As an example, getVolumeStore calls processGlobalFlags, and getComposer calls getVolumeStore, which means each compose commands call processGlobalFlags at least twice (one in getVolumeStore, one in its own command if global flags are used).
Either way, we should highlight that helper functions (e.g., getVolumeStore) should be refactored to not call processGlobalFlags directly and instead use the struct created by the command.
nice catch. I have update all the PR to make sure call processGlobalFlags once for each command
@AkihiroSuda PTAL
|
2025-04-01T06:38:16.442041
| 2023-05-20T18:10:48
|
1718245291
|
{
"authors": [
"cgwalters"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4911",
"repo": "containers/bootc",
"url": "https://github.com/containers/bootc/issues/101"
}
|
gharchive/issue
|
drop dependency on xfs_freeze
Reported via email; we should instead just directly invoke the ioctl. ostree has code for this we could make into a public C API, including the complex dance of doing this by forking a helper process to be robust against interruption.
This was done in https://github.com/containers/bootc/pull/102/commits/15de723ac2a9707523967aa4ee5c29875b94baa8
|
2025-04-01T06:38:16.488033
| 2024-07-31T12:58:02
|
2440000300
|
{
"authors": [
"codecov-commenter",
"kwilczynski",
"rphillips",
"saschagrunert"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4912",
"repo": "containers/conmon-rs",
"url": "https://github.com/containers/conmon-rs/pull/2350"
}
|
gharchive/pull-request
|
WIP: Update nixpkgs
What type of PR is this?
/kind dependency-change
What this PR does / why we need it:
Bump nixpkgs.
Which issue(s) this PR fixes:
None
Special notes for your reviewer:
None
Does this PR introduce a user-facing change?
None
:warning: Please install the to ensure uploads and comments are reliably processed by Codecov.
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 35.75%. Comparing base (4e0f474) to head (2444a51).
Report is 632 commits behind head on main.
:exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality.
Additional details and impacted files
@@ Coverage Diff @@
## main #2350 +/- ##
==========================================
- Coverage 37.53% 35.75% -1.78%
==========================================
Files 15 15
Lines 1268 1264 -4
Branches 414 420 +6
==========================================
- Hits 476 452 -24
- Misses 526 552 +26
+ Partials 266 260 -6
@kwilczynski PTAL
/approve
/lgtm
/lgtm
|
2025-04-01T06:38:16.492014
| 2022-07-25T12:54:11
|
1316796519
|
{
"authors": [
"rphillips",
"saschagrunert"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4913",
"repo": "containers/conmon-rs",
"url": "https://github.com/containers/conmon-rs/pull/552"
}
|
gharchive/pull-request
|
Use --systemd-cgroup for runtime commands
What type of PR is this?
/kind bug
What this PR does / why we need it:
This fixes the passing of systemd related annotations to the config, for example: https://github.com/opencontainers/runc/pull/2224
Which issue(s) this PR fixes:
None
Special notes for your reviewer:
None
Does this PR introduce a user-facing change?
Using `--systemd-cgroup` for runtime commands.
PTAL @rphillips @haircommander
lgtm
I'll defer to @haircommander for the final lgtm
Peter is AFK. Going to lgtm and we can iterate if we would like to change this PR further.
/lgtm
miss click
/lgtm
|
2025-04-01T06:38:16.494486
| 2022-09-12T10:04:08
|
1369610672
|
{
"authors": [
"haircommander",
"saschagrunert"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4914",
"repo": "containers/conmon-rs",
"url": "https://github.com/containers/conmon-rs/pull/703"
}
|
gharchive/pull-request
|
Add critest run to CI
What type of PR is this?
/kind ci
What this PR does / why we need it:
We now run a critest suite with a pinned CRI-O release for each PR.
Which issue(s) this PR fixes:
None
Special notes for your reviewer:
Needs https://github.com/containers/conmon-rs/pull/702 to not timeout during the attach test.
Does this PR introduce a user-facing change?
None
/lgtm
|
2025-04-01T06:38:16.506632
| 2021-12-07T08:27:10
|
1073061562
|
{
"authors": [
"saschagrunert"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4915",
"repo": "containers/containrs",
"url": "https://github.com/containers/containrs/pull/705"
}
|
gharchive/pull-request
|
Do not upload tarball any more
What type of PR is this?
/kind failing-test
What this PR does / why we need it:
There is no need for the artifact right now se we can stop uploading it every commit.
Which issue(s) this PR fixes:
None
Special notes for your reviewer:
None
Does this PR introduce a user-facing change?
None
Merging to unblock the CI
|
2025-04-01T06:38:16.612376
| 2024-10-11T11:27:35
|
2581138179
|
{
"authors": [
"TristanCacqueray",
"ericcurtin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4916",
"repo": "containers/ramalama",
"url": "https://github.com/containers/ramalama/issues/285"
}
|
gharchive/issue
|
Serve command fails silently
Hello,
Running the serve command with a wrong model name outputs a container ID that dies instantly without a reason.
Here is an example reproducer:
$ ramalama serve huggingface://ibm-granite/granite-7b-instruct
6aa6dc6cca217b28656990f69a7c602c28c40f64289d8cdb5c55e6ecccdc0828
$ ramalama ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
$
I tried to add an helpful message by calling the pull command before serve, but it looks like there is another issue in the glob f"{args.store}/models/*/{model}" because the Model class doesn't have a str method, resulting in the path to never match because it looks like:
ramalama/models/*/<ramalama.huggingface.Huggingface object at 0x7fa1c6e659a0>
How would you recommend to fix that issue?
I kinda wish we didn't detach by default because of reasons like this
I meant to add a check before calling ramalama.run_container in the Process CLI entrypoint, to ensure the model is pulled before starting the service.
The problem with just adding a check for the file, is that only helps debug one very specific issue. Like it is already checked if the file exists just in llama-server and we can't see the output
Right, so not detaching by default sounds great to me, may I propose such a change?
|
2025-04-01T06:38:16.618057
| 2017-09-25T19:41:08
|
260391923
|
{
"authors": [
"coveralls",
"grahamwhaley",
"jodh-intel",
"mcastelino",
"sameo"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4917",
"repo": "containers/virtcontainers",
"url": "https://github.com/containers/virtcontainers/pull/384"
}
|
gharchive/pull-request
|
WIP: Huge Page Support: Add support for huge pages (needs qemu revendor)
Add support to launch virtual machines where the RAM is
allocated using huge pages. This is useful for running
with a user mode networking stack, and for custom setups
which require high performance and low latency.
Note: Needs revendoring of ciao/qemu changes https://github.com/ciao-project/ciao/pull/1449
Signed-off-by: Manohar Castelino<EMAIL_ADDRESS>
lgtm
just to be clear - this is 'pre-allocated' with huge pages is it?
no matter...
lgtm
@mcastelino Please rebase, the needed qemu changes should be there now.
LGTM
Coverage decreased (-0.01%) to 65.928% when pulling 27902cf28573d310ebe192bb682f5da8dbf02207 on mcastelino:topic/huge_pages into 04e519d758c64d52b51670fac68d58f2ddb752c7 on containers:master.
|
2025-04-01T06:38:16.621576
| 2022-01-14T21:33:03
|
1104142109
|
{
"authors": [
"Furisto",
"codecov-commenter",
"utam0k"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4918",
"repo": "containers/youki",
"url": "https://github.com/containers/youki/pull/613"
}
|
gharchive/pull-request
|
Ensure youki runs under podman
We need to be more defensive when deleting a container as podman executes delete twice and cleans up the cgroup of the container before we have a chance to clean it up.
Fixes #607
Codecov Report
Merging #613 (d0d756a) into main (52f83a0) will decrease coverage by 0.03%.
The diff coverage is 0.00%.
@@ Coverage Diff @@
## main #613 +/- ##
==========================================
- Coverage 70.04% 70.01% -0.04%
==========================================
Files 82 82
Lines 10924 10933 +9
==========================================
+ Hits 7652 7655 +3
- Misses 3272 3278 +6
Wow, I was surprised to know that.
podman executes delete twice
LGTM
|
2025-04-01T06:38:16.628890
| 2019-10-01T08:20:39
|
500738297
|
{
"authors": [
"bgranvea",
"dduportal",
"mmatur"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4919",
"repo": "containous/traefik",
"url": "https://github.com/containous/traefik/issues/5557"
}
|
gharchive/issue
|
Self-generated certificate is regenerated every 15 seconds
Do you want to request a feature or report a bug?
Bug
Did you try using a 1.7.x configuration for the version 2.0?
[ ] Yes
[X] No
What did you do?
I configured Traefik with Docker provider and swarmMode=true, and I deployed a service with an HTTPS entrypoint and no tls options.
What did you expect to see?
Traefik should generate a certificate for HTTPS and keep it, so that it can be added to trusted certificates in a browser.
What did you see instead?
Traefik generates a new certificate for HTTPS every 15 seconds, so the browser keeps telling that the certificate is not trusted.
Output of traefik version: (What version of Traefik are you using?)
Version: 2.0.1
Codename: montdor
Go version: go1.13.1
Built: 2019-09-26T16:18:03Z
OS/Arch: linux/amd64
If applicable, please paste the log output in DEBUG level (--log.level=DEBUG switch)
time="2019-09-30T13:54:52Z" level=debug msg="No default certificate, generating one"
...
time="2019-09-30T13:55:07Z" level=debug msg="No default certificate, generating one"
...
time="2019-09-30T13:55:22Z" level=debug msg="No default certificate, generating one"
Hi @bgranvea ,
Thanks for your interest in the project.
It seems to be a duplicate of #5381
So I will close this issue.
hum yes maybe it is the same root cause as #5381 but with swarmMode=true, it seems even worse as the certificate is regenerated even when no new container is deployed.
And I don't understand why you don't consider #5381 as a bug: IMHO the self-generated certificate feature is not usable with a browser.
Hi @bgranvea, for the "swarmMode" making this behavior even worse, it's related to the issues #5419. This issue had been closed and will be in the next 2.0.x release. You can already test it with the image containous/traefik:experimental-v2.0 if you want to live on the edge :)
As for #5381, we consider the "self generated certificate" as a "degraded" mode. If you're willing to use a self signed certificate for all of your services, we recommend generating one and setting Traefik to use it as explained in https://docs.traefik.io/v2.0/https/tls/. I personnaly use the awesome tool mkcert https://github.com/FiloSottile/mkcert which generate for multiple domains and wildcards including *.local, and install the CA in my browsers.
|
2025-04-01T06:38:16.646950
| 2019-10-01T20:54:47
|
501141004
|
{
"authors": [
"dduportal",
"fragtom",
"valentinvieriu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4920",
"repo": "containous/traefik",
"url": "https://github.com/containous/traefik/issues/5563"
}
|
gharchive/issue
|
docker compose traefik2.0 not work with other docker services
Do you want to request a feature or report a bug?
Bug
Did you try using a 1.7.x configuration for the version 2.0?
[ ] Yes
[x] No
What did you do?
I setup docker-compose.yaml file with docker example whoami service. on a second docker-compose I added service whoami2.
What did you expect to see?
Firstly, expect to see routings on dashboard and services. Second expectation: see response on whoami.docker.localhost (works) and reponse on url whoami2.docker.localhost.
What did you see instead?
for bot services see information on dashboard. For whoami2.docker.localhost got "Gateway Timeout"
Output of traefik version: (What version of Traefik are you using?)
/ # traefik version
Version: 2.0.1
Codename: montdor
Go version: go1.13.1
Built: 2019-09-26T16:18:03Z
OS/Arch: linux/amd64
What is your environment & configuration (arguments, toml, provider, platform, ...)?
## docker-compose.yaml
version: '3.7'
services:
traefik:
image: "traefik:v2.0.1"
container_name: traefik
ports:
- "80:80" # The HTTP port
- "8080:8080" # The Web UI (enabled by --api)
command:
- "--log.level=DEBUG"
volumes:
- /var/run/docker.sock:/var/run/docker.sock # So that Traefik can listen to the Docker events
- ./traefik.yaml:/etc/traefik/traefik.yaml
whoami:
# A container that exposes an API to show its IP address
image: containous/whoami
labels:
- "traefik.http.routers.whoami.rule=Host(`whoami.docker.localhost`)"
## traefik.yml
# Docker configuration backend
providers:
docker:
exposedbydefault: true
entryPoints:
web:
# Listen on port 80 for incoming requests
address: :80
# API and dashboard configuration
api:
insecure: true
# other-folder/docker-compose.yaml
version: '3.7'
services:
whoami2:
image: containous/whoami
labels:
- "traefik.http.routers.whoami2.rule=Host(`whoami2.docker.localhost`)"
If applicable, please paste the log output in DEBUG level (--log.level=DEBUG switch)
docker up traefik whoami
Creating network "traefik_default" with the default driver
Creating traefik_whoami_1 ... done
Creating traefik ... done
Attaching to traefik_whoami_1, traefik
whoami_1 | Starting up on port 80
traefik | time="2019-10-01T20:51:55Z" level=info msg="Configuration loaded from file: /etc/traefik/traefik.yaml"
docker up traefik whoami2
Starting docker-wordpress_whoami2_1 ... done
Attaching to docker-wordpress_whoami2_1
whoami2_1 | Starting up on port 80
output whoami
curl -H Host:whoami.docker.localhost http://<IP_ADDRESS>
Hostname: 6b4f2dfe944f
IP: <IP_ADDRESS>
IP: <IP_ADDRESS>
GET / HTTP/1.1
Host: whoami.docker.localhost
User-Agent: curl/7.63.0
Accept: */*
Accept-Encoding: gzip
X-Forwarded-For: <IP_ADDRESS>
X-Forwarded-Host: whoami.docker.localhost
X-Forwarded-Port: 80
X-Forwarded-Proto: http
X-Forwarded-Server: 9f14dd78222b
X-Real-Ip: <IP_ADDRESS>
output whoami2
curl -H Host:whoami2.docker.localhost http://<IP_ADDRESS>
Gateway Timeout
Hi @fragtom , thanks for your interest in the project and this complete report!
As this is a configuration issue, could you switch to our community forum at https://community.containo.us/ to get support and knowledge from the community (maintainers included) please? The reason is that we use Github for tracking bugs and feature requests.
As a hint for your next steps (if it doesn't work use a community topic to mention me), you might want to check this page from Docker documentation at https://docs.docker.com/compose/networking/.
Each docker-compose stack has it's own private network: Traefik cannot reach the whoami containers because in another network. You have to merge the stacks OR declare external networks (https://docs.docker.com/compose/compose-file/#external-1).
@fragtom the way I solve this I create a traefik network in the traefik's docker-compose file:
version: '3.7'
networks:
traefik:
name: traefik
services:
traefik:
image: "traefik:v2.0.1"
container_name: 'traefik'
networks:
- traefik
ports:
- "80:80" # The HTTP port
- "8080:8080" # The Web UI (enabled by --api)
command:
- "--log.level=DEBUG"
volumes:
- /var/run/docker.sock:/var/run/docker.sock # So that Traefik can listen to the Docker events
- ./traefik.yaml:/etc/traefik/traefik.yaml
And then use the same network in the other docker-compose files:
# other-folder/docker-compose.yaml
version: '3.7'
networks:
traefik:
external: true
services:
whoami2:
image: containous/whoami
container_name: 'whoami'
networks:
- 'traefik'
labels:
- "traefik.http.routers.whoami2.rule=Host(`whoami2.docker.localhost`)"
Please notice the external part.
Hope this helps
That works for me, thnks
On Fri 25. Oct 2019 at 08:33, Valentin Vieriu<EMAIL_ADDRESS>wrote:
@fragtom https://github.com/fragtom the way I solve this I create a
traefik network in the traefik's docker-compose file:
version: '3.7'
networks:
traefik:
name: traefik
services:
traefik:
image: "traefik:v2.0.1"
container_name: 'traefik'
networks:
- traefik
ports:
- "80:80" # The HTTP port
- "8080:8080" # The Web UI (enabled by --api)
command:
- "--log.level=DEBUG"
volumes:
- /var/run/docker.sock:/var/run/docker.sock # So that Traefik can listen to the Docker events
- ./traefik.yaml:/etc/traefik/traefik.yaml
And then use the same network in the other docker-compose files:
other-folder/docker-compose.yaml
version: '3.7'
networks:
traefik:
external: true
services:
whoami2:
image: containous/whoami
container_name: 'whoami'
networks:
- 'traefik'
labels:
- "traefik.http.routers.whoami2.rule=Host(whoami2.docker.localhost)"
Please notice the external part.
Hope this helps
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/containous/traefik/issues/5563?email_source=notifications&email_token=AAEPGIGPC6YMIWQ72KKATVTQQKHK3A5CNFSM4I4OXPX2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOECHK4LA#issuecomment-546221612,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AAEPGIG3RPNWTJLPFQZK7N3QQKHK3ANCNFSM4I4OXPXQ
.
--
--
Tom Stark
Hochäckerstr. 115
81737 München
Tel.: 01520 911 9008
--
|
2025-04-01T06:38:16.682645
| 2024-04-23T15:57:19
|
2259235762
|
{
"authors": [
"jlsherrill",
"swadeley"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4921",
"repo": "content-services/content-sources-backend",
"url": "https://github.com/content-services/content-sources-backend/pull/645"
}
|
gharchive/pull-request
|
Refs 2422: Add pending task metrics
Summary
Adds new metrics:
* Average queue time of pending task
* Length in queue for the oldest pending task
* Number of pending tasks
Also:
* Increases default metric calculations to 30 seconds from 5
* makes interval for metric calculation configurable
* adds log level override for metrics so you can set the rest of the app to trace while metrics are set to debugging
Testing steps
on a new db:
make run
in another tab:
make repos-import
go run cmd/external-repos/main.go nightly-jobs
This will kick off a ton of tasks, you can monitor with:
curl localhost:9000/metrics | grep task_stats
This will update every ~30 seconds.
Checklist
[ ] Tested with snapshotting feature disabled and pulp server URL not configured if appropriate
note this is currently built ontop of https://github.com/content-services/content-sources-backend/pull/637
will rebase once merged.
https://issues.redhat.com/browse/HMS-2422
/retest
/retest
/retest
added a test to check for queued_at updates.
Also i realized that the 'number of days' until the cdn cert expires was only calculated at startup, so i changed it to be calculated as part of the go routine that updates the metrics. At startup it was set to zero, so to keep that from firing an alert, i set it to calculate that one metric at startup (and only that metric, because others require the db to be up, which may not actually have happened yet?)
added
|
2025-04-01T06:38:16.694030
| 2022-06-24T22:25:27
|
1284244730
|
{
"authors": [
"Andrewgdewar",
"jlsherrill"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4922",
"repo": "content-services/content-sources-frontend",
"url": "https://github.com/content-services/content-sources-frontend/pull/9"
}
|
gharchive/pull-request
|
Fixes 41: Add create content modal
To test this PR, one will need to have checked out this backend PR (or have it merged to master).
This adds the bulk create modal.
This adds the following required API's:
Arch/Version repo
Validate Content
Bulk Create
Additional changes:
Added empty state/emptyTable state/Loading table state with skeleton
Worked entirely too much on gpgKey integration (80% functionality)
Hid said GPG key functionality.
Replaced @rverdile's react Context code (sorry!) to avoid this bug
Updated Eslint rules
Updated some cloud-services npm modules to latest
Added formik/Yup node modules
Fixed an issue with optimistic updates.
Added "Add 20" magic button within the create modal.
Fixed issue when deleting last item on a page > 1, where the page wouldn't navigate the user back.
Fixed issue with page not resetting when changing the search filters
Still needs:
Add readable specific field errors if bulk create fails (currently just throws generic), waiting on this ticket
Direction on what to do when we have an url that fails it's head request (currently prevents continuing):
Note on the above (not sure if this is considered a bug), pinging an address that has an http redirect will return success.
Example: https://quay.io/stuffandthings/thatprobablydontexist
https://issues.redhat.com/browse/HMSCONTENT-41
I'm guessing this appeared after hiding the gpg key section, but now the dropdown for versions clips the end of the section and you can't see it:
One issue:
enter a url without any version info in it
select some set of versions
go back to the url and change it slightly
versions change back to 'any version'
I thought we had mentioned this and you said it shouldn't do this, but i may be misremembering, i know you said it was tricky
seeing a weird format issue when trying to add a repo and the versions were auto-selected from the url. The structure of the distribution_versions is wrong:
[
{
"name": "test4",
"url": "http://yum.theforeman.org/katello/4.2/katello/el7/x86_64/",
"distribution_arch": "x86_64",
"distribution_versions": [
[
"el7"
]
],
"gpgKey": ""
},
{
"name": "test8",
"url": "http://mirror.centos.org/centos/8-strea/BaseOS/x86_64/os/",
"distribution_arch": "x86_64",
"distribution_versions": [
"el9"
],
"gpgKey": ""
}
]
The first object in this list was auto-selected, the 2nd one was not. Notice that there is an array within an array in the first object which isn't valid
seeing a weird format issue when trying to add a repo and the versions were auto-selected from the url. The structure of the distribution_versions is wrong:
[
{
"name": "test4",
"url": "http://yum.theforeman.org/katello/4.2/katello/el7/x86_64/",
"distribution_arch": "x86_64",
"distribution_versions": [
[
"el7"
]
],
"gpgKey": ""
},
{
"name": "test8",
"url": "http://mirror.centos.org/centos/8-strea/BaseOS/x86_64/os/",
"distribution_arch": "x86_64",
"distribution_versions": [
"el9"
],
"gpgKey": ""
}
]
The first object in this list was auto-selected, the 2nd one was not. Notice that there is an array within an array in the first object which isn't valid
Working on this now. I attempted an incomplete fix yesterday.
|
2025-04-01T06:38:16.711274
| 2024-09-26T06:53:13
|
2549679580
|
{
"authors": [
"shayan-NECX",
"ttbuffey"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4923",
"repo": "contentauth/c2pa-python-example",
"url": "https://github.com/contentauth/c2pa-python-example/issues/10"
}
|
gharchive/issue
|
Saving image loses c2pa information
Hi, thanks for the work. I am trying to save image with c2pa information in app.py code, it does save the image but then when i read c2pa information, it says there is no manifest attached.
However, when i follow the tutorial, as is, and do the curl request, then the output image saved by it, has the c2pa information.
I dont understand why would this happen? Why saving image after receiving image bytes, loses the c2pa content ? Please guide.
@app.route("/attach", methods=["POST"])
def resize():
request_data = request.get_data()
manifest = json.dumps({
"title": "image2.jpg",
"format": "image/jpeg",
"claim_generator_info": [
{
"name": "c2pa test",
"version": "0.0.1"
}
],
"assertions": [
{
"label": "c2pa.actions",
"data": {
"actions": [
{
"action": "c2pa.edited",
"softwareAgent": {
"name": "C2PA Python Example",
"version": "0.1.0"
}
}
]
}
}
]
})
builder = Builder(manifest)
signer = create_signer(sign, SigningAlg.ES256,
cert_chain, "http://timestamp.digicert.com")
result = io.BytesIO(b"")
builder.sign(signer, "image/jpeg", io.BytesIO(request_data), result)
with open('result.txt','w') as f:
print(result.getvalue(),file=f)
image = Image.open(result)
# Save the image IO object to a file
image.save('output_image5.jpg')
image = Image.open(io.BytesIO(result.getvalue()))
# # Save the image bytes to a file
image.save('output_image6.jpg')
with open('byte_result.txt','w') as f:
print(result.getvalue(),file=f)
print("End")
return result.getvalue()
@shayan-NECX it seems you run the api successful at the first place. however i get some issue that I need your help to confirm where is wrong ehen i follow the guide. I will close the issue.
Regards,
Shayan
From: ttbuffey @.>
Sent: Tuesday, October 15, 2024 7:48 AM
To: contentauth/c2pa-python-example @.>
Cc: Bhatti, Shayan Ali @.>; Mention @.>
Subject: Re: [contentauth/c2pa-python-example] Saving image loses c2pa information (Issue #10)
You don't often get email from @.*** Learn why this is importanthttps://aka.ms/LearnAboutSenderIdentification
@shayan-NECXhttps://github.com/shayan-NECX it seems you run the api successful at the first place. however i get some issue that I need your help to confirm where is wrong ehen i follow the guide. I can't even run the curl successfully.
i also get someissues, after run the curl command I get an error when run api.py
builder.sign(signer, "image/jpeg", io.BytesIO(request_data), result) File "/Users/i302707/Documents/pii/c2pa-python-example/ven_3.10.5/lib/python3.10/site-packages/c2pa/c2pa_api/c2pa_api.py", line 129, in sign return super().sign(signer, format, C2paStream(input), C2paStream(output)) File "/Users/i302707/Documents/pii/c2pa-python-example/ven_3.10.5/lib/python3.10/site-packages/c2pa/c2pa/c2pa.py", line 925, in sign rust_call_with_error( File "/Users/i302707/Documents/pii/c2pa-python-example/ven_3.10.5/lib/python3.10/site-packages/c2pa/c2pa/c2pa.py", line 283, in rust_call_with_error uniffi_check_call_status(error_ffi_converter, call_status) File "/Users/i302707/Documents/pii/c2pa-python-example/ven_3.10.5/lib/python3.10/site-packages/c2pa/c2pa/c2pa.py", line 313, in uniffi_check_call_status raise error_ffi_converter.lift(call_status.error_buf) c2pa.c2pa.c2pa.Error.Signature: reason='COSE error parsing certificate
what i have done is first run
python setup.py create-key-and-csr 'CN=John Smith,O=C2PA Python Demo'
kms-signing.csr file is cgenerated
then create fake rootCA
openssl req -x509 \ -days 1825 \ -newkey rsa:2048 \ -keyout rootCA.key \ -out rootCA.crt
then sign the CSR created in step 1 with the temporary test CA key
openssl x509 -req \ -CA rootCA.crt \ -CAkey rootCA.key \ -in kms-signing.csr \ -out kms-signing.crt \ -days 365 \ -copy_extensions copyall
then create chain.pem
cat kms-signing.crt rootCA.crt > chain.pem
FLASK_KMS_KEY_ID="$KMS_KEY_ID" FLASK_CERT_CHAIN_PATH="./chain.pem" flask run
curl command, i get the "COSE error parsing certificate" error from the builder.sign
—
Reply to this email directly, view it on GitHubhttps://github.com/contentauth/c2pa-python-example/issues/10#issuecomment-2413824139, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AYBQUZGWZHEZPLGG5VREOELZ3UFLRAVCNFSM6AAAAABO4JOWEOVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDIMJTHAZDIMJTHE.
You are receiving this because you were mentioned.Message ID: @.***>
|
2025-04-01T06:38:16.720070
| 2020-06-09T10:45:04
|
635320884
|
{
"authors": [
"FlorianReimold",
"Kerstin-Keller",
"NewProggie",
"gocarlos",
"photex"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0000.json.gz:4924",
"repo": "continental/ecal",
"url": "https://github.com/continental/ecal/issues/54"
}
|
gharchive/issue
|
Regarding Conan dependencies.
The current conanfile.txt in this repo is using packages that are not automatically or easily resolved.
It would be helpful to know where kwc_bintray is and add this to the README so that others can use Conan.
Digging around the commit and issue history I see that this was contributed by someone from Kai Wolf (so kwc seems to make sense). So it appears as though they have a self-hosted artifactory service but not all the packages were listed. At which point I realized that it's probably https://bintray.com/kwc which is rather easy in hindsight.
Running conan remote add kwc_bintray https://api.bintray.com/conan/kwc/conan resolved this for me.
Then you apparently also need bincrafters:
conan remote add bincrafters https://api.bintray.com/conan/bincrafters/public-conan
But you finally end up with the issue of the version of protobuf being requested is not apparently available in the bincrafters public conan repository. However, the requested version is available on conan-center. But still missing is protoc_installer.
It would be useful to add some documentation for the level of Conan support in the repo and how to use it.
Hi @photex,
I totally agree that Conan integration is not smooth yet. We worked together with @NewProggie to the Conan integration.
Unlike the behavior you get from Conan Center packages, we want the Conan integration to be truly optional. E.g. only the cmake_paths generator should be used. (e.g. all target names used in CMake should be identical to those that you get by pure CMake installs).
This is why many packages were repackaged instead of using the official packages.
However, the current recommendation to build eCAL is to use submodule integration for thirdparty dependencies for an easy build on Windows, and on Linux use packages provided by the system package manager when available (e.g. on Ubuntu 18.04, hdf5, protobuf, qt, … can all be retrieved by apt-get).
More experienced users can use any way to provide the necessary dependencies.
But it still makes sense to put a chapter in the Readme file on Conan support / integration.
However, the requested version is available on conan-center. But still missing is protoc_installer.
Conan Center decided to package protoc and protobuf separately, mainly to facilitate cross-builds. With the kwc packages, they are packaged together (which I think they should be. Since they belong together, they are always required in the same version).
Good catch. I think there should be another section in the README.md containing:
# Install Conan
$ pip install --upgrade pip
$ pip install --upgrade conan
$ conan config set general.revisions_enabled=True
# Add Conan remotes with pre-compiled dependencies
$ conan remote add -f kwc_bintray https://api.bintray.com/conan/kwc/conan
$ conan remote add -f conan-center https://bintray.com/conan/conan-center
$ conan remote add -f bincrafters https://api.bintray.com/conan/bincrafters/public-conan
Let me quickly add another PR for this.
Thanks for the explanations.
@NewProggie neither bincrafters or conan-center had protoc_installer for the specified version when I tried it today. Can you verify that this works with these remotes?
actually the packages in the conan center are now being reworked to allow transparent conan usage, e.g. same target names of the packages when installed and found via cmake...
if some package is still not work this way, its a bug in the package that needs fixing...
Closing this discussion-thread due to inactivity
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.