added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:10:15.158089
| 2024-02-13T19:05:01
|
2132968222
|
{
"authors": [
"ayokunle321",
"chris-vanderveen",
"samad-zeeshan"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13699",
"repo": "CMPUT301W24T08/DroidDesign",
"url": "https://github.com/CMPUT301W24T08/DroidDesign/issues/32"
}
|
gharchive/issue
|
US 02.02.03
As an attendee, I want to update information such as name, homepage, and contact information on my profile.
Acceptance Criteria
[ ] Attendees can easily access their profile settings to edit personal information.
[ ] Specific editable fields for name, homepage URL, and contact information (e.g., email, phone number) are clearly editable.
[ ] The system validates input for correctness and completeness, ensuring URLs are valid and contact information follows the correct format.
[ ] Changes are saved and reflected immediately upon submission.
[ ] Attendees receive confirmation that their profile has been successfully updated.
[ ] Clear error messages are displayed for invalid inputs, guiding users to correct the information.
Created the Story board and User Interface Mock ups for this!
Completed
Completed
|
2025-04-01T04:10:15.202147
| 2017-12-07T20:37:02
|
280270569
|
{
"authors": [
"FlexionSonarQube",
"saquino0827"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13700",
"repo": "CMSgov/qpp-conversion-tool",
"url": "https://github.com/CMSgov/qpp-conversion-tool/pull/488"
}
|
gharchive/pull-request
|
CPC+ Endpoint 2
Information
QPPCT-502
Changes proposed in this PR:
New endpoint to retrieve the CPC+ XML File by file Id
Checklist
[x] All JUnit tests pass (mvn clean verify).
[x] New unit tests written to cover new functionality.
[x] Added and updated JavaDocs for non-test classes and methods.
[x] No local design debt. Do you feel that something is "ugly" after your changes?
[x] Updated documentation (README.md, etc.) depending if the changes require it.
SonarQube analysis reported 5 issues
5 major
Watch the comments in this conversation to review them.
SonarQube analysis reported 1 issue
1 major
Watch the comments in this conversation to review them.
SonarQube analysis reported 1 issue
1 minor
Watch the comments in this conversation to review them.
SonarQube analysis reported 1 issue
1 major
Watch the comments in this conversation to review them.
SonarQube analysis reported 7 issues
3 major
4 minor
Watch the comments in this conversation to review them.
SonarQube analysis reported 2 issues
2 minor
Watch the comments in this conversation to review them.
|
2025-04-01T04:10:15.271086
| 2020-06-29T23:56:22
|
647744539
|
{
"authors": [
"aidanheerdegen",
"angus-g"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13701",
"repo": "COSIMA/cosima-cookbook",
"url": "https://github.com/COSIMA/cosima-cookbook/issues/185"
}
|
gharchive/issue
|
Expose more information in API calls
It would be good to return more information from the get_experiments and get_variables methods. This would help with data discovery as it would allow simple GUIs like this
https://github.com/COSIMA/cosima-cookbook/issues/102#issuecomment-651428383
to filter on more attributes that are stored in the database.
I'd be for returning as much information as possible, or having a flag to do so.
What about providing the widget in the cookbook itself? Then it would have access to all the database fields needed.
Yeah, but you can never make something that is perfect for everything, and someone else might want to make their own widget, so doesn't it make sense to just make the API capable of supporting that use case?
What is the down side? Too large a table?
I guess they're not mutually exclusive. I was just thinking that the cookbook is the right level to implement a widget so people wouldn't have to make their own (plus you can exploit querying/relations at the database level)
The tables are pretty unwieldy as-is, so there's no harm in adding more info to them if the intention is to consume them with further filters!
I agree a widget should be part of the cookbook ... eventually. I did think about pulling some of the code out and rolling my own method to return data, but the nice thing about getting that from the cookbook API is that I don't have to do that, and I can prototype stuff in a notebook, and once it is working just stick it straight in the cookbook. Which is sort of the point of an API.
When I said "too large a table" I meant the pandas table that is returned from the query, no the SQL DB table. Sorry for the ambiguity.
Makes sense.
I meant the pandas table that is returned from the query,
That's what I was referring to as well. For most experiments these tables are way too big to just look at, so you have to further process them anyway, so I don't see why they can't have more info in them.
|
2025-04-01T04:10:15.279836
| 2020-04-26T19:14:05
|
607090365
|
{
"authors": [
"SomeMoosery",
"tesla809"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13702",
"repo": "COVID-19-electronic-health-system/Corona-tracker",
"url": "https://github.com/COVID-19-electronic-health-system/Corona-tracker/issues/649"
}
|
gharchive/issue
|
[DOCS] Add twitter button to all github repos
โ ๏ธ IMPORTANT: Please fill out this template to give us as much information as possible to consider/implement this update.
Summary
Adding Twitter button on top of our GitHub will help increase our social media.
see:
https://gist.github.com/studiopress/6867036
Sample code from Babel's README:
See how Babel did it:
https://github.com/babel/babel
Motivation
Why are we doing this?
More follower on twitter. More legitimacy and integration across our social channels. Social capital from twitter can lend itself to our project and vice versa. This will help our marketing and engagement.
What use cases does it support?
This will help our marketing and engagement.
What is the expected outcome?
We get more followers on twitter which helps our organic social media
Possible Alternatives
We can have a twitter link inside the README.md, but that is less clear.
Having a prominent button helps us get attention and a higher click rate to sign up.
Additional Context
Sample code from Babel's README:
See how Babel did it:
https://github.com/babel/babel
Should we add it to all github repos @whoabuddy ?
I'd say this should be moved to the .github org as I think this should be present across all repos
|
2025-04-01T04:10:15.280798
| 2020-03-16T03:16:15
|
581967764
|
{
"authors": [
"sedmo",
"tesla809"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13703",
"repo": "COVID-19-electronic-health-system/Corona-tracker",
"url": "https://github.com/COVID-19-electronic-health-system/Corona-tracker/pull/41"
}
|
gharchive/pull-request
|
Theme palettehotfix
the color and background properties was duplicated for themePalette.css so I created seperate classes for background , slider , and text
I pair programmed with you on this and looks good. Merged.
|
2025-04-01T04:10:15.303772
| 2018-02-20T04:19:18
|
298467470
|
{
"authors": [
"scheng98",
"tqshao"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13704",
"repo": "CPFL/Autoware",
"url": "https://github.com/CPFL/Autoware/issues/1097"
}
|
gharchive/issue
|
Dependency warnings during installation
Hello develoeprs,
I am installing Autoware using the kinetic docker.
During installation, I noticed there were quite some warnings like the following:
[91mCMake Warning at /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'yaml-cpp' but neither 'yaml-cpp_INCLUDE_DIRS' nor 'yaml-cpp_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package) sensing/drivers/lidar/packages/velodyne/velodyne_pointcloud/CMakeLists.txt:39 (catkin_package)
[0m[91mCMake Warning at /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'python-yaml' but neither 'python-yaml_INCLUDE_DIRS' nor 'python-yaml_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package) sensing/drivers/lidar/packages/velodyne/velodyne_pointcloud/CMakeLists.txt:39 (catkin_package)
[91mCMake Warning at /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'common' but neither 'common_INCLUDE_DIRS' nor 'common_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package) computing/perception/localization/lib/fast_pcl/ndt_cpu/CMakeLists.txt:16 (catkin_package)
[91mCMake Warning at /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'system_lib' but neither 'system_lib_INCLUDE_DIRS' nor 'system_lib_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/kinetic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package)
Will this cause any problem when I run Autoware? What can I do to eliminate these errors? Thanks for your time!
try this one,
catkin_make -Dcv_bridge_DIR=/usr/share/cv_bridge/cmake/ <<<< replace it with your cv_bridge path
|
2025-04-01T04:10:15.313137
| 2024-08-28T13:36:18
|
2492105368
|
{
"authors": [
"codecov-commenter",
"croyzor"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13705",
"repo": "CQCL/guppylang",
"url": "https://github.com/CQCL/guppylang/pull/423"
}
|
gharchive/pull-request
|
chore: Update hugr llvm and test new op lowering
Update hugr_llvm so that we get lowering for multiply and divide. Test the new functionality
Drive-by: Fix inkwell to version 0.4.0 - the new release (0.5.0) has just come out and has api breaking changes
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 92.48%. Comparing base (d5a1b6f) to head (3d5907c).
Additional details and impacted files
@@ Coverage Diff @@
## main #423 +/- ##
=======================================
Coverage 92.48% 92.48%
=======================================
Files 48 48
Lines 5178 5178
=======================================
Hits 4789 4789
Misses 389 389
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
2025-04-01T04:10:15.314048
| 2018-04-24T21:59:04
|
317410489
|
{
"authors": [
"jaeddy",
"vthorsson"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13706",
"repo": "CRI-iAtlas/shiny-iatlas",
"url": "https://github.com/CRI-iAtlas/shiny-iatlas/issues/85"
}
|
gharchive/issue
|
Specify T Cell Receptor Module
Identify what would go into a dedicated T Cell Receptor Repertoire Module. Include Scott Brown and possibly others in the discussion.
Moving to internal discussion at CRI-iAtlas/iatlas-planning#1.
|
2025-04-01T04:10:15.327882
| 2018-10-31T03:51:07
|
375787309
|
{
"authors": [
"FongYuan",
"coveralls"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13707",
"repo": "CS2103-AY1819S1-F10-2/main",
"url": "https://github.com/CS2103-AY1819S1-F10-2/main/pull/157"
}
|
gharchive/pull-request
|
Add colours to loan and bike lists
The list of loans in the UI is now yellow, and the list of bikes (accessible via listbikes) is blue.
This has been done to differentiate the list of loans from the list of bikes, so the user can more easily notice which list they are looking at.
Pull Request Test Coverage Report for Build 547
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 91.713%
Totals
Change from base Build 526:
0.0%
Covered Lines:
1981
Relevant Lines:
2160
๐ - Coveralls
|
2025-04-01T04:10:15.335246
| 2017-11-11T04:24:41
|
273117901
|
{
"authors": [
"April0616"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13708",
"repo": "CS2103AUG2017-W09-B1/main",
"url": "https://github.com/CS2103AUG2017-W09-B1/main/issues/199"
}
|
gharchive/issue
|
Delete Command response msg
Show the names of all the deleted persons first, followed by their details
Need to specify which tag the user deleted
fix by #216
|
2025-04-01T04:10:15.345727
| 2021-06-10T14:48:13
|
917432040
|
{
"authors": [
"jayhsu0627",
"yajain"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13709",
"repo": "CSAILVision/semantic-segmentation-pytorch",
"url": "https://github.com/CSAILVision/semantic-segmentation-pytorch/issues/261"
}
|
gharchive/issue
|
How to get learning rate to decrease more slowly?
How should I change the value of lr_pow in order to make my learning rate decrease more slowly. Currently the lr_pow is at a default value of 0.9.
By using --lr_pow', default=0.9, type=float, help='power in poly to drop LR'
|
2025-04-01T04:10:15.354311
| 2024-04-06T12:39:46
|
2229225581
|
{
"authors": [
"Charlie-XIAO",
"ROMEEZHOU"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13710",
"repo": "CSCI-SHU-410-SE-Project/Deskulpt",
"url": "https://github.com/CSCI-SHU-410-SE-Project/Deskulpt/pull/10"
}
|
gharchive/pull-request
|
TST unit tests for bundler
I have written some unit tests for bundler, which cover the following cases:
bundling a simple file that doesn't explicitly use React
bundling a simple file that explicitly uses React
bundling a file that imports from a .js file using relative path with extension, e.g. import { getMessage } from "./utils.js";,
bundling a file that imports from a .js file using relative path without extension
bundling a file that imports from a directory using relative path
bundling a file that imports from a .jsx file using relative path without extension, in this case, both the file being imported and the main file has const React = window.__DESKULPT__.defaultDeps.React;
bundling a file that imports from a .jsx file, the file being imported doesn't define React
bundling a file that imports from a .jsx file ,the main file doesn't define React
bundling a file that imports using an absolute path (which should raise errors)
bundling a file that imports from a file that is beyond the root (which should raise errors)
You can cd to src-tauri and run the test by
cargo test
You may also check the coverage visualization by running
cargo llvm-cov --open
@CSCI-SHU-410-SE-Project/core-dev Please let me know if there are any problems :)
LGTM, thanks @ROMEEZHOU! We shall take care of code coverage and more error cases in follow-up PRs.
|
2025-04-01T04:10:15.355117
| 2022-03-15T12:08:49
|
1169587765
|
{
"authors": [
"krzyzanowskim",
"rvsrvs"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13711",
"repo": "CSCIX65G/SwiftCrossCompilers",
"url": "https://github.com/CSCIX65G/SwiftCrossCompilers/pull/17"
}
|
gharchive/pull-request
|
Swift 5.6 Configuration
Swift 5.6 Configuration. This one uses Apple provided arm64 tarball ๐
dayum... how cool is that?
|
2025-04-01T04:10:15.364120
| 2023-11-27T13:32:36
|
2012282839
|
{
"authors": [
"calebevans"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13712",
"repo": "CSPI-QE/firewatch",
"url": "https://github.com/CSPI-QE/firewatch/issues/111"
}
|
gharchive/issue
|
[Maintenance] Make the release v2 changes available in OpenShift CI
What needs to be done:
[x] Make sure the image will be available
[x] Update the firewatch report issues ref to use the new arguments (replaced underscores with hyphens)
[x] Add a new boolean environment variable for the --report-success command
[x] Add new environment variables for the "default" values
[x] Fix any config syntax issues that will be introduced with these changes
This PR is where all of this work is held: https://github.com/openshift/release/pull/46282
|
2025-04-01T04:10:15.393025
| 2024-11-25T09:30:07
|
2689948260
|
{
"authors": [
"AlbinoGeek",
"CTNOriginals"
],
"license": "BSD-4-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13713",
"repo": "CTN-Originals/RoleMentionCooldown",
"url": "https://github.com/CTN-Originals/RoleMentionCooldown/pull/4"
}
|
gharchive/pull-request
|
Massive Cleanup
Linted all the markdown files (this makes renderers+git happy)
New lines at end of file (this makes git happy)
Removed hundreds of unused imports
Sorted imports (this makes git happy)
Work in progress
this needs to be pushed to the development branch, not to stable.
headsup: The readme has been updated with #5
The rebase broke a few things, there are three files out of sync I have to correct. I MIGHT restart these changes on a fresh branch, not sure yet. Will update later.
Yeah a restart might be better here as there have been a good few new commits.
|
2025-04-01T04:10:15.423972
| 2023-05-25T19:57:32
|
1726410402
|
{
"authors": [
"Adamits",
"kylebgorman"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13714",
"repo": "CUNY-CL/yoyodyne",
"url": "https://github.com/CUNY-CL/yoyodyne/issues/66"
}
|
gharchive/issue
|
Updates and Documentation for WandB sweeps
Based on other work were doing, we should add some documentation and make necessary tweaks for running a W&B sweep with this codebase.
Add documentation and examples of running WandB sweeps with Yoyodyne.
Make updates to codebase so PTL and WandB play nice wrt logging hyperparameters, etc.
Update PTL to log max validation accuracy.
Let me also add that the documentat should probably show how to retrieve best runs from the wandb API too.
I guess relatedly it would also be nice to have a system for easily pointing W&B run id's to yoyodyne logging, etc.
Working on this now. Was wondering if you think we should add train args so that it can be called in such a way that a wandb agent trains from a sweep (by adding a wandb_sweep_id and max_num_runs arg), or if this should be a seperate scripts that we maintain in the library (something like train_wandb_agent.py).
Other notes:
I was not able to find anything on how to log the max validation accuracy in PTL, and let it propogate that logging to wandb, so instead I just do wandb.define_metric('val_accuracy', summary='max') when wandb logging is enabled.
PTL tries to log the model hparams to the wandb when the WandbLogger is enabled, causing a warning, because they also get logged when we start the sweep agent. See here: https://github.com/wandb/wandb/issues/2641. I do not know how to fix this, since it does not seem to be PTL behavior I can toggle, and we need the PTL WandbLogger in order to also log runtime metrics. I think we can just let it happen it for now?
Working on this now. Was wondering if you think we should add train args so that it can be called in such a way that a wandb agent trains from a sweep (by adding a wandb_sweep_id and max_num_runs arg), or if this should be a seperate scripts that we maintain in the library (something like train_wandb_agent.py).
While I'm not sure I have enough context to get this yet, I think I am fine just including docs and a sample script for doing wandb stuff. It's hard for me to imagine doing this effectively using yoyodyne-train alone, I guess? I assume you did your sweeping using custom Python, right?
I was not able to find anything on how to log the max validation accuracy in PTL, and let it propogate that logging to wandb, so instead I just do wandb.define_metric('val_accuracy', summary='max') when wandb logging is enabled.
SGTM.
PTL tries to log the model hparams to the wandb run when the WandbLogger is enabled, causing a warning, because they also get logged when we start the sweep agent. See here: [CLI] wandb: WARNING Config item 'hyperparam_name' was locked by 'sweep' (ignored update)ย wandb/wandb#2641. I do not know how to fix this, since it does not seem to be PTL behavior I can toggle, and we need the PTL WandbLogger in order to also log runtime metrics. I think we can just let it happen it for now?
Let's just suppress the warning in __init__.py then, and add a TODO to investigate this at the PTL level later.
Working on this now. Was wondering if you think we should add train args so that it can be called in such a way that a wandb agent trains from a sweep (by adding a wandb_sweep_id and max_num_runs arg), or if this should be a separate script that we maintain in the library (something like train_wandb_agent.py).
While I'm not sure I have enough context to get this yet, I think I am fine just including docs and a sample script for doing wandb stuff. It's hard for me to imagine doing this effectively using yoyodyne-train alone, I guess? I assume you did your sweeping using custom Python, right?
Yeah I just have a train_wandb_agent.py script that calls the functions in train.py. So do we need a directory at the top-level of our repository called examples or similar? Or do you think its better to have train_wandb_agent.py live alongside train.py?
Let's just suppress the warning in __init__.py then, and add a TODO to investigate this at the PTL level later.
Sounds good!
Working on this now. Was wondering if you think we should add train args so that it can be called in such a way that a wandb agent trains from a sweep (by adding a wandb_sweep_id and max_num_runs arg), or if this should be a separate script that we maintain in the library (something like train_wandb_agent.py).
While I'm not sure I have enough context to get this yet, I think I am fine just including docs and a sample script for doing wandb stuff. It's hard for me to imagine doing this effectively using yoyodyne-train alone, I guess? I assume you did your sweeping using custom Python, right?
Yeah I just have a train_wandb_agent.py script that calls the functions in train.py. So do we need a directory at the top-level of our repository called examples or similar? Or do you think its better to have train_wandb_agent.py live alongside train.py?
Yes that's what I'd suggest. I'd have one for running the sweep and, optionally, one for grabbing the results from W&B.
I don't know if we need to modify the project file to register the existence of that directory, but prevent it from being installed as part of the package...something to look out for: browse the verbose installation info and you should see what happens there.
@kylebgorman Should we leave this open until we've played with the examples and are sure the scripts are sufficient, and documentation is good enough?
I was thinking I'd commit it and then just take it for a spin next...
Okay, sure. I'd like to take it for a spin first.
Sorry, I just meant this issue -- not the PR!
Sorry, I just meant this issue -- not the PR!
Got it, yea I was confused at first.
|
2025-04-01T04:10:15.428174
| 2017-02-02T14:17:56
|
204886104
|
{
"authors": [
"Jwhiles",
"des-des"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13715",
"repo": "CYPIAPT-LNDSE/friendly-friends",
"url": "https://github.com/CYPIAPT-LNDSE/friendly-friends/pull/34"
}
|
gharchive/pull-request
|
Fix build issue
Quite small changes, this means we can successfuly build, and then removes our build file from gitignore so we can deploy the build.
You might only want them commited on your ghpages branch ..
Sorry, will remove the gitignore changes and repush
|
2025-04-01T04:10:15.442188
| 2024-10-07T22:30:56
|
2571632913
|
{
"authors": [
"CZEMacLeod",
"julealgon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13716",
"repo": "CZEMacLeod/MSBuild.SDK.SystemWeb",
"url": "https://github.com/CZEMacLeod/MSBuild.SDK.SystemWeb/issues/83"
}
|
gharchive/issue
|
Nested Web.config files are not treated as Content for publishing and transforming
As I switch our old WebForms projects to leverage publishing, I noticed that only the root web.config files are properly considered for deployment, transformed and copied on publish, while web.config files inside of other folders in the project (used usually for setting Location-based permissions) are completely ignored.
I believe it should be possible to treat these nested web.config files the same way as the root one so that they properly participate in the publishing pipeline.
I've not yet worked around this issue, so suggestions on how to do so besides manually marking each individual file as Content in Visual Studio are welcome.
You probably want something like
<ItemGroup>
<Content Include="**\Web.config" Exclude="Web.config" />
</ItemGroup>
This is one of the advantages of the SDK style project.
I would always use a location tag in the root web.config, instead of one in each folder, as otherwise you may end up with odd behavior as not all information is inherited. There is some special stuff in place for razor views where you need a customized web.config in the Views folder already in the Razor SDK.
<location path="some/plugin/location/handler.ashx">
<system.web>
<httpRuntime maxRequestLength="1048576" executionTimeout="600" />
</system.web>
<system.webServer>
<security>
<requestFiltering>
<!-- 500MB in bytes, default is 30000000 or approx. 28.6102 Mb-->
<requestLimits maxAllowedContentLength="1073741824" />
</requestFiltering>
</security>
</system.webServer>
</location>
<location path="some/plugin2/location/otherhandler.ashx">
<system.web>
<pages validateRequest="false" />
<httpRuntime requestValidationMode="2.0" executionTimeout="600" />
</system.web>
</location>
You probably want something like
<ItemGroup>
<Content Include="**\Web.config" Exclude="Web.config" />
</ItemGroup>
This appears to have done the trick, thanks!
Do you know what would be needed to also support transformations on these non-root web.config files? I assume just the above is not enough?
I would always use a location tag in the root web.config, instead of one in each folder, as otherwise you may end up with odd behavior as not all information is inherited.
Thanks for the advice, but this would not scale well for this particular project. This is a legacy WebForms project with over a thousand pages in a multitude of folders, and we have permission management everywhere. If we concentrated all of them in the main file (or even if it was isolated in a dedicated file linked via configSource, it would be completely unmaintainable, especially as this project has multiple feature teams handling it at the same time each with their own pages and folders.
Having native support for these nested web.config files would be really nice for this use case.
Do you know what would be needed to also support transformations on these non-root web.config files? I assume just the above is not enough?
Are you talking about including publish profile transforms for msdeploy, or configuration transforms like debug / release etc?
The section of the SDK that deals with this is https://github.com/CZEMacLeod/MSBuild.SDK.SystemWeb/blob/add70be345fd8a988c3a659bd9b482a6988087fd/src/MSBuild.SDK.SystemWeb/Sdk/MSBuild.SDK.SystemWeb.DefaultItems.props#L9-L24
You might be able to something along the lines of
<ItemGroup Condition="'$(EnableWebFormsDefaultItems)'=='true'">
<Content Include="**\Web.config" />
<None Include="@(_WebConfigConfigurations->'**\Web.%(Identity).config')">
<DependentUpon>%(RelativeDir)\Web.config</DependentUpon>
</None>
<Content Include="**\Web.*.config" Exclude="@(None)" />
</ItemGroup>
Thanks for the advice, but this would not scale well for this particular project. This is a legacy WebForms project with over a thousand pages in a multitude of folders, and we have permission management everywhere. If we concentrated all of them in the main file (or even if it was isolated in a dedicated file linked via configSource, it would be completely unmaintainable, especially as this project has multiple feature teams handling it at the same time each with their own pages and folders.
Fair enough. As I said - it should work - but I've had issues with this approach in non-sdk projects, let alone with the hacks the SDK does. If you get it to work for you, all the better.
Having native support for these nested web.config files would be really nice for this use case.
You might need to tweak the DependentUpon line above, but refer to https://learn.microsoft.com/en-us/visualstudio/msbuild/msbuild-well-known-item-metadata?view=vs-2022 if what I gave doesn't work.
You might need to adjust the line to be something like
<None Include="@(_WebConfigConfigurations->'**\Web.%(Identity).config')" Condition="EXISTS('@(_WebConfigConfigurations->'**\Web.%(Identity).config')'"> to work around #67 or you could implement something like the fix in #68
Do you know what would be needed to also support transformations on these non-root web.config files? I assume just the above is not enough?
Are you talking about including publish profile transforms for msdeploy, or configuration transforms like debug / release etc? The section of the SDK that deals with this is
https://github.com/CZEMacLeod/MSBuild.SDK.SystemWeb/blob/add70be345fd8a988c3a659bd9b482a6988087fd/src/MSBuild.SDK.SystemWeb/Sdk/MSBuild.SDK.SystemWeb.DefaultItems.props#L9-L24
You might be able to something along the lines of
<ItemGroup Condition="'$(EnableWebFormsDefaultItems)'=='true'">
<Content Include="**\Web.config" />
<None Include="@(_WebConfigConfigurations->'**\Web.%(Identity).config')">
<DependentUpon>%(RelativeDir)\Web.config</DependentUpon>
</None>
<Content Include="**\Web.*.config" Exclude="@(None)" />
</ItemGroup>
I was referring to the normal transforms, using Web.Debug.config/Web.Release.config etc. We are currently not doing any XML transforms on non-root web.config files, but I know they are supposed to work on the legacy project format. If someone could implement that in this SDK it would be more complete/closer to the full capabilities of the legacy project.
Perhaps your suggestions above could help someone implementing it directly in the SDK.
Your initial suggestion unblocked me to get our publish pipeline going, but I added it to our project with a note in hopes that this is implemented as part of the SDK in the future and we can then remove the customization:
<ItemGroup>
<!--
HACK:
Adds all nested `web.config` files manually. Ideally, this should be handled by the SDK, but it only
considers the root `web.config` file at the moment.
More information here:
- https://github.com/CZEMacLeod/MSBuild.SDK.SystemWeb/issues/83
-->
<Content Include="**\Web.config" Exclude="Web.config" />
</ItemGroup>
Thankfully we've only had to add this to a single project thus far out of the 8 projects we have currently using this SDK.
|
2025-04-01T04:10:15.449231
| 2016-02-18T17:35:30
|
134644548
|
{
"authors": [
"SRGDamia1"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13717",
"repo": "CZOData/Search.CriticalZone.org",
"url": "https://github.com/CZOData/Search.CriticalZone.org/issues/4"
}
|
gharchive/issue
|
#7 Expose webservices (like WaterOneFlow) as links in dataset results
Just as links to "display files" and "metadata xml" files are separately listed on right of each search result.
Also make those webservices a "collection" in the resources.
If the web service links were readily found at search.criticalzone.org then they would be a great example for the BiG-CZ tools/iPython notebooks.
[--- Commented from Asana.com
#commenter Sara Damiano
---[aa]
As the display file is parsed, add a link to the web service link to the metadata.
[--- Commented from Asana.com
#commenter Sara Damiano
---[aa]
Right now the standard metadata does not include links to the webservices. Both we might be able to just assume that if it's a parsed display file then it is available as a web service via the individual CZO's Hydroserver.
[--- Commented from Asana.com
#commenter Sara Damiano
---[aa]
These services would be exposed as "link types" for each metadata record. Web services should be recorded at a granular level, such as CUAHSI WaterOneFlow and OGC WMS, not as a lumped "Data Web Service". An outcome of this is that any given metadata record on search.criticalzone.org may have several link types, including: "Display Files", "CZO Dataset" "WaterOneFlow", "WMS", and so on.
[--- Commented from Asana.com
#commenter Emilio Mayorga
---[aa]
|
2025-04-01T04:10:15.453045
| 2022-03-21T01:34:13
|
1174778098
|
{
"authors": [
"Caaz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13718",
"repo": "Caaz/rimworld-typhon",
"url": "https://github.com/Caaz/rimworld-typhon/issues/86"
}
|
gharchive/issue
|
Add Greater Mimics
Description
Greater mimics, are mimics, except spikier and beefier. They can take more hits and deal more damage. Will probably require nerfing normal mimics to make the strength noticeable.
Technical stuff
Greater mimics should be a chance spawn from mimic multiplying.
The succ
Greater mimics should have an ability to one-shot face-suck kill people. Might be a bit brutal to players -- worth giving them a warning, and/or enabling this behavior only when difficulty settings are above a certain level.
Art Needed
A spikier mimic texture essentially
Audio Needed
Can probably re-use mimic audio.
Wound up working on this myself, this may wind up becoming a thing fairly soon.
|
2025-04-01T04:10:15.454057
| 2020-10-26T20:29:22
|
729882581
|
{
"authors": [
"TheGreatCabbage",
"felixonmars"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13719",
"repo": "CabbageDevelopment/qasync",
"url": "https://github.com/CabbageDevelopment/qasync/issues/13"
}
|
gharchive/issue
|
Missing git tag for 0.9.5 release
It would be nice to keep PyPI releases and git tags in sync :)
Good point, done :+1:
|
2025-04-01T04:10:15.498566
| 2018-02-11T18:55:27
|
296214896
|
{
"authors": [
"CaffeineViking"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13720",
"repo": "CaffeineViking/osgw",
"url": "https://github.com/CaffeineViking/osgw/issues/3"
}
|
gharchive/issue
|
Add noise/bump mapping for small wave ripples
It would be a waste to create a very detailed mesh for the Gestner wave, when small details can be done in the fragment shader with normal/bump mapping. Perhaps we don't even need to use Gerstner waves in this case, and can use a pre-defined normal map to create the ripples.
Implemented in the latest branch in master.
|
2025-04-01T04:10:15.499566
| 2023-01-23T14:12:36
|
1553172823
|
{
"authors": [
"Cai1Hsu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13721",
"repo": "Cai1Hsu/osu",
"url": "https://github.com/Cai1Hsu/osu/pull/1"
}
|
gharchive/pull-request
|
Add separate sensitivity support for android
Requires:
[x] Cai1Hsu/osu-framework/pull/1
needs tests on android devices to merge
|
2025-04-01T04:10:15.515562
| 2017-11-17T15:43:26
|
274906073
|
{
"authors": [
"mstreit",
"sgratzl"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13722",
"repo": "Caleydo/lineupjs",
"url": "https://github.com/Caleydo/lineupjs/issues/463"
}
|
gharchive/issue
|
group selection status not in sync with 'select all'
Release number or git hash: develop
Web browser version and OS: chrome
Steps to reproduce
stratify by some attribute
aggregate first group
checking 'select all' checkbox
Observed behavior
Group is unaffected by select all status (out of sync)
Expected behavior
Checking 'select all' will select all groups and items
|
2025-04-01T04:10:15.524555
| 2021-01-27T14:51:14
|
795168977
|
{
"authors": [
"TomSmithCGAT",
"csdaw"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13723",
"repo": "CambridgeCentreForProteomics/camprotR",
"url": "https://github.com/CambridgeCentreForProteomics/camprotR/issues/16"
}
|
gharchive/issue
|
Move from Travis to GitHub actions
Unfortunately @TomSmithCGAT, after all your wrangling with Travis we'll need to move to GitHub actions as Travis may no longer be free to use (see this blog post amongst others).
Fortunately, this seems easy enough and there are simple instructions in that blog post. I'm happy to do this myself when I get the chance.
Can this be closed?
I was going to close it when {ts}_test_test_test is merged with master.
|
2025-04-01T04:10:15.581711
| 2015-05-13T16:06:23
|
76036999
|
{
"authors": [
"tayzlor"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13724",
"repo": "Capgemini/Apollo",
"url": "https://github.com/Capgemini/Apollo/issues/180"
}
|
gharchive/issue
|
Improve weave spec tests
At the moment we only test the bridge. We could -
test for the container
test docker options for weave
test the network interfaces file
Closing via https://github.com/Capgemini/Apollo/pull/208
|
2025-04-01T04:10:15.586343
| 2021-05-24T03:12:52
|
899201888
|
{
"authors": [
"Nixinova"
],
"license": "ISC",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13725",
"repo": "CarbonChess/CarbonSite",
"url": "https://github.com/CarbonChess/CarbonSite/issues/14"
}
|
gharchive/issue
|
Add chess variants
Variants like antichess, king of the hill, three-check, horde, crazy house, atomic, etc.
Would be useful for FenFurnace to have an interface to customise piece movement ability
https://github.com/CarbonChess/FenFurnace/issues/5
|
2025-04-01T04:10:15.588615
| 2023-09-05T15:56:09
|
1882251669
|
{
"authors": [
"EpikCloudFR"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13726",
"repo": "Carbubu/upptime",
"url": "https://github.com/Carbubu/upptime/issues/89"
}
|
gharchive/issue
|
โ ๏ธ Carbubu has degraded performance
In 17a7260, Carbubu (https://carbubu.fr) experienced degraded performance:
HTTP code: 200
Response time: 5220 ms
Resolved: Carbubu performance has improved in b21c1cc after 13 minutes.
|
2025-04-01T04:10:15.602452
| 2023-10-19T10:09:48
|
1951797362
|
{
"authors": [
"NumericalMax",
"matt74"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13727",
"repo": "CardioKit/PeakWatch",
"url": "https://github.com/CardioKit/PeakWatch/issues/31"
}
|
gharchive/issue
|
problem with package dependency
Hello, I tried your sample code but I have issue with peakswift package dependency. See the screenshot. I have an error when I want to add it to the project. I'm using latest Xcode 15.0
Hi Matthias, thank you for bringing this issue to our attention. We have now changed how the PeakSwift package fetches submodules, i.e., using https instead of ssh. May you check whether the package includes the submodules correctly and PeakWatch compiles. Best Max
Thanks for the fix. It's working well now.
|
2025-04-01T04:10:15.625548
| 2023-07-25T00:17:53
|
1819324551
|
{
"authors": [
"CarliJoy",
"DiscordLiz",
"JelleZijlstra",
"NeilGirdhar",
"erictraut",
"mikeshardmind"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13728",
"repo": "CarliJoy/intersection_examples",
"url": "https://github.com/CarliJoy/intersection_examples/issues/6"
}
|
gharchive/issue
|
TypedDict rules are unsound
I am looking at the current rules in https://github.com/CarliJoy/intersection_examples/blob/main/specification.rst#typeddicts, which say that if two TypedDict members of an intersection share the same field with different types, then the intersection contains a field that is a union of those two types.
from typing import TypedDict
class A(TypedDict):
a: int
class B(TypedDict):
a: str
def takes_a(a: A) -> None:
print(a["a"] + 1)
def takes_ab(ab: A & B) -> None:
assert_type(ab["a"], int | str)
takes_a(ab) # legal, A & B is compatible with A
b: B = {"a": "x"}
takes_ab(b) # legal, but throws an error at runtime
A similar hole:
def takes_ab2(ab: A & B) -> None:
ab["a"] = 42 # legal, as ab["a"] is of type int | str
b2: B = {"a": "x"}
takes_ab2(b2)
print(b2["a"]) # now it's an int!
Possibly the type of ab["a"] should actually be int & str, not int | str.
As I mentioned in other threads, I don't think the Intersection PEP should make any mention of protocols (structural types), nominal types, callables, overloads, etc. The same applies to TypedDict (which happens to be a structural type). If we define the rules for intersections correctly, there will be no need to mention any of these.
Consider that there was never a need to describe how a TypedDict works with a union; the behavior is obvious from the basic rules of how unions work. Likewise, there should be no need to describe how a TypeDict works with intersections.
I'm going to refrain from reviewing the draft PEP spec until it's a bit further along and incorporates some of the feedback from these threads. In its current form, it has some significant problems. The sections on "Handling Callables" and "Protocol Reduction" contain many statements that are demonstrably incorrect (as Jelle points out above by example). I think these need to be replaced by a more general set of simple rules like the ones I suggest in this comment.
@erictraut I just started with quite a naive understanding of typing and no expert at all. I would be very happy if you could change the specification to a more sound definition.
I can give you access to the repo once I am home (don't have my security key with me at the moment) or you simply create a pull request.
I do not agree that we don't need special mentioning of TypedDict and other Collections.
For instance the current implementation of Intersection of basedmypy does not handle TypedDict correctly, it simply makes it a dict.
This kind of collections are special and it should be included in the Specification that they need to be handled in order to have a common understanding for them within all type checkers.
Do I understand it correctly that properties of function that mismatch should be intersections and not unions?
I do not agree that we don't need special mentioning of TypedDict and other Collections. ... This kind of collections are special
Why are they special? They're types within the type system. The type system also supports collection types besides TypedDict, and there will likely be new collection types added in the future. If we define intersections in terms of TypedDict or any other specific type, we would need to define how every newly-added type works with intersections. I'd consider that a failure. We need to define the rules of intersection in a way that work with all types in the type system โ including all types that exist today and all types that might be defined in the future.
It's fine to use some specific examples (e.g. that involve TypedDict) within the PEP to demonstrate how these rules work when applied to actual types, but I think it's very problematic if the rules themselves include mention of TypedDict or any other type in the type system that obeys the rules of subtyping.
For instance the current implementation of Intersection of basedmypy does not handle TypedDict correctly
If we define intersections correctly, then it will be obvious that the basedmypy implementation is broken in this regard. There's no need to describe in the PEP how intersections work with every type as long as the rules can be applied to all types. Type checkers already require a lot of special casing when implementing certain operations on types like TypedDict, enums, properties, and tuples, but the rules for intersection should apply to all of these in the same way.
Do I understand it correctly that properties of classes/types that mismatch should be intersections and not unions?
By "properties", I assume you mean "attributes". (Sorry to nitpick, but the term "property" has a specific meaning in Python, and I don't think it applies here.) We need to define the rule for how attribute accesses work when applied to intersections. It's already defined for unions today; we simply need to extend that to intersections. And yes, if two types support the same attribute, then the result of accessing that attribute from an intersection of those two types will result in an intersection.
class A:
a: Sequence[int]
b: Sequence[int]
class B:
a: datetime
def func(x: A | B, y: A & B):
reveal_type(x.a) # Sequence[int] | datetime
reveal_type(y.a) # Sequence[int] & datetime
reveal_type(x.b) # Error
reveal_type(y.b) # Sequence[int]
reveal_type(x.c) # Error
reveal_type(y.c) # Error
if we "merge" a protocol/typeddics we actually apply the same rules to each property?
That depends on what you mean by "merge". I'm not sure that a "merge" operation needs to be defined. For example, there's no place in pyright today where I attempt to "merge" two types.
That depends on what you mean by "merge". I'm not sure that a "merge" operation needs to be defined. For example, there's no place in pyright today where I attempt to "merge" two types.
Okay than it is probably my lack in understanding of typings/typer checkers.
How to do implement/define the access of attributes when subtyping (is this the correct term)?
Given the examples you gave it is probably really possible to minimize the specification a lot.
I just don't have time for that within the next two weeks.
How to do implement/define the access of attributes when subtyping (is this the correct term)?
If we go with something entirely generalizable, the types of attributes of an intersection generalize as
(A & B).foo = (&) (A & B โ foo).foo
Put another way, "Taking the types in the intersection which have a member foo and then taking the intersection of each member foo found this way"
As an exhaustive pattern for example
(A & B).foo
if A provides foo and B doesn't, A.foo
If B provides foo and A doesn't, B.foo
if both A and B provide foo, A.foo & B.foo
if neither A nor B provide foo, Never (Empty set)
Note that this is one of the arguments against reducing Any & T
Any & T
T provides foo: Any.foo & T.foo (T.foo?)
T doesn't provide foo: Any.foo (Any)
(A & B).foo = (&) {foo โ A & B}.foo
Do you have a formal proof of this I can take a look at? Intuitively, I think it should be correct and if it is, it puts a neat bow around Any as no longer needing to be treated as special, but I'm not sure what this would mean in a few cases.
class A:
def foo(self) -> None:
...
class B:
@staticmethod
def foo() -> None:
...
class C:
@classmethod
def foo(cls) -> None:
...
x: A & B & C = ... # lets just ignore that I don't know how to make such a class in a sensical way
x.foo() # this is safe, and Known to be `None` ?
x: Callable[..., None] & Callable[[None], None]
x() # None ?
These are the easy cases, now lets see:
Given the rule you provided
x: tuple[int, ...] & tuple[str, ...]
x[0] # int & str? So this is `Never` then because you can't subclass both?
No formal proof yet, I didn't have as much time to work on this today, but it seemed like a more general rule that applied even to the "Special" case of Any (and later to other special cases of gradual typing that were shown)
All of what you have shown there is how I would interpret the rule applying, including the last one being an impossibility, yes. However, I would say that in the last case, x could still be an empty tuple, as it was possible to specify the tuple, but then impossible to assign elements
Actually, I have a useful case for that specification already. This could be used for the not-uncommon request to mark args or kwargs as not taken in paramspec, by making it impossible to provide them in a way which satisfies the type of them.
I've thought about this for a while, but I don't know if I agree with either the conflicting tuples or the conflicting classes. Can you explain why this definition should work with or without a formal proof of it?
Sure.
First of all, if this is the only definition and only constraint of those classes A, B, and C, I think a Union is actually more appropriate.
What this essentially is doing is saying that if something meets the definition of A and if A has a definition of foo, whatever just met the definition of A has a definition of foo that can be used as a replacement for A's definition of foo. (ie. compatible subclasses of A are possible, along with A)
This should be completely uncontroversial I think, as it is how singular type specifications already work.
If we have already established that something meets the definitions for A, of B, and of C, then we only need to check if interactions with it are consistent with how we would interact with A, B, and C.
For the record, right now the A & B & C case with A.foo as a method, B.foo as a staticmethod, and C.foo as a classmethod, each taking no arguments other than what is bound by descriptors, I don't think it is possible to create a class consistent with all 3, so we would fail at the assignment of an instance to the definition, not at access to foo, but lets ignore that because this shows how we do not necessarily need to check for all kinds of consistency at every use, and this is helpful when considering special cases such as Any, allowing us to check consistency at specific points in time in the ways they matter at that time.
We have the above definition that does not care if that is possible or not and which can still determine if the use of an attribute is consistent with the definitions. This leaves anything based on concrete types to be checked only at assignment while checking attributes by compatibility with each use, rather than some symbolic idea compositing the entire type.
This also means that type checkers do not need to create and store some virtual subclass to check attributes. They can lazily construct an internal representation to check only the attributes which are used after initial assignability concerns are handled.
so in the case of tuple[int, ...] & tuple[str, ...] it is essentially tuple[int & str, ...] because that would be consistent with __getitem__ for each.
Because
ti: tuple[int, ...] = (,)
ts: tuple[str, ...] = (,)
are both valid, we have a tuple that can handle this in terms of assignability, and that cannot ever have a length greater than 0, because it would be impossible for any tuple to be constructed where __getitem__ could return an item satisfying both.
allowing us to check consistency at specific points in time in the ways they matter at that time.
Oh, okay, so this explains why I thought it shouldn't work with that rule and how it could, especially when you clarified that you didn't think it was possible to even get to that point without error in the case of the classes.
It's an interesting approach that does appear to work.
I've removed the section on type-dicts as per the discussion above.
|
2025-04-01T04:10:15.627725
| 2024-06-25T14:49:17
|
2372900698
|
{
"authors": [
"danoswaltCL",
"ppratikcr7"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13729",
"repo": "CarnegieLearningWeb/UpGrade",
"url": "https://github.com/CarnegieLearningWeb/UpGrade/issues/1694"
}
|
gharchive/issue
|
Metrics Delete Modal, type 'delete'
Instead of typing in the whole path of a metric to delete it, the modal should ask for them to type 'delete' like we do elsewhere
Also update the message on modal from "Type the metric path separated by space to confirm deletion:" to "Type 'delete' to confirm deletion:".
Use the env text "global.delete-confirmation.message.text" from en.json
|
2025-04-01T04:10:15.657284
| 2018-03-05T18:12:32
|
302405511
|
{
"authors": [
"AdriSolid",
"davidmanzanares",
"jorgesancha",
"makella",
"nerik"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13730",
"repo": "CartoDB/carto-vl",
"url": "https://github.com/CartoDB/carto-vl/issues/127"
}
|
gharchive/issue
|
Heatmap support
[x] Custom PoCs: NY foot trafic dataset, gecat dataset and WWI_ships dataset
[x] Support for grid-alignment (NY dataset)
[x] Bi-cubic interpolation
[x] Isolines PoC
[ ] Support non-square windows
[ ] API
[ ] Support for limitting heatmap grid size based on zoom
[ ] Workarounds lack of OES_texture_float and OES_texture_float_linear. Important for mobile.
[ ] Automatic discovery of min/max values
API pseudo Proposal
Create a heatmap by accumulating the count with the SUM function and use 0 as the low value for the ramp, and 100 for the high value.
Each heatmap cell will cover 20x20 pixels
heatmap: ramp(linear(cellSum($count), 0, 100), tealrose)
heatmapResolution: 20
Implementation proposal
The rendering part has already been more or less addressed.
Regarding the API, we probably won't support every feature from the start, but, I think this API is achievable and flexible. For example, it allows setting different aggregation functions. Internally, those aggregations could be sent to Maps API, greatly reducing MVT size.
It also allows to specify the MIN/MAX values for the ramp, which at first could be delegated to the user, but, it could be set dynamically in the future, like:
heatmap: ramp(linear(
cellSum($count),
viewportMin(cellSum($count)),
viewportMax(cellSum($count))
),
tealrose)
Where viewportMin(...) returns the minimum value of the cells on the viewport. However, having a constant expression, like a constant number or a zoom-dependant expression is probably fine for most cases.
We could have isopleth maps with:
heatmap: ramp(buckets(cellSum($count),
0, 20, 50, 100
),
tealrose)
And isolines with:
heatmap: ramp(isolines(cellSum($count),
0, 20,50,100, 0.1
),
tealrose)
??? I don't like this one
...
cc @rochoa
@davidmanzanares
Above, you mention:
"Create a heatmap by accumulating the count with the SUM function and use 0 as the low value for the ramp, and 100 for the high value."
Just to make sure, we are after more than just a visual representation of density, right? In order to get deeper into the map types that we should support, I think I need to understand the basics of what is being implemented a little better.
In addition, depending on what the data are (crime vs. GPS vs. weather) the user will likely want more control on the methods used and the visual output of the interpolated surface.
I'm not an expert in "heat maps" but to go beyond visual density, we need:
cell size (20 x 20)
the search radius for the density calculation (do we have this?)
the density calculations/interpolation methods (count,min,max,avg... what else? IDW,... )
In addition, I'm curious:
how we handle this through zoom
we need better CARTOColor schemes to map to the range (0-100) likely going from "cool" to "hot" with a good, multi-hue change in between.
I think once I understand what our goal is and how we will be implementing it, that I can contribute to the conversation better!
In addition, aside from enhanced CARTOColor schemes, if a user defines their own colors, we will likely need to think about another color interpolation space as cielab provides a uniform transition without introducing additional hues... and in the case of these interpolated surfaces, that will be key.
this is interesting
https://www.movebank.org/node/6400
Something new about it? Thanks!
@davidmanzanares
@makella
@rochoa
Hey there. Just a quick note, we are just starting to evaluate different solutions for a new version of our animated heatmap at GlobalFishingWatch.
So far our potential options are deck.gl, CARTO VL or go with writing our own MGL custom layer. (in case you're interested this is our current approach)
Very excited by that gif posted by @davidmanzanares . This feature could really be a dealmaker for us. Is it still on roadmap?
Hi @nerik Thanks for the feedback! And really interesting post.
We have not forgotten about this and it is definitely coming, but unfortunately maybe not fast enough. It is unlikely we get to this before end of Q2. Hope that gives you a frame of reference
Hi @jorgesancha. Thanks for the reply, and noted for the timeline.
One question though: there's a gif attached to the issue, does that mean you have some working proof-of-concept somewhere? If possible we'd be interested in checking it out.
I believe this is the branch. Bear in mind that this way behind the master branch, based in code that no longer exists and probably with a lot of hacks just to proof the concept
Yes, there were multiple branches. The last one I think it is square-grid, but it also contains some unwanted changes to the sources.
I think that a merge would be too difficult at this point too. Rewriting the functionality while reading the diff of the branch is probably the easiest way.
You can read the diff here: https://github.com/CartoDB/carto-vl/compare/square-grid
The interesting parts that could be copied are on renderer.js and on the shaders.
Keep in mind that the branch never made a public API to configure the heatmap.
|
2025-04-01T04:10:15.659604
| 2015-06-29T08:12:19
|
91732603
|
{
"authors": [
"jsanz",
"nobuti"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13731",
"repo": "CartoDB/cartodb.js",
"url": "https://github.com/CartoDB/cartodb.js/issues/547"
}
|
gharchive/issue
|
Handling 404 error on createLayer/createVis
When entering a wrong URL on createLayer/createVis the on('error') event is not fired. I guess this has to be related with using JSONP and a dirty workaround is possible using timeouts (example). I'm not sure if we can do anything about this but it'd be nice if wrong urls could be handled by the library.
There has been no activity on this issue for several months and that's why are closing it (automatically). If you think this still needs to be addressed please reopen it. The good news is that we are now working on the next version of CARTO.js, stay tuned.
|
2025-04-01T04:10:15.662466
| 2015-09-15T15:27:08
|
106581200
|
{
"authors": [
"alonsogarciapablo",
"iriberri"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13732",
"repo": "CartoDB/cartodb.js",
"url": "https://github.com/CartoDB/cartodb.js/issues/681"
}
|
gharchive/issue
|
CartoDB.js 3.15.5: Uncaught TypeError: Cannot set property '_attributions' of undefined
If a CartoDB is used over a Leaflet map in which the attributionsControl (more info here) is set to false:
map = L.map("map", {
zoom: 11,
center: [28.175278, -81.370102],
layers: [ggl2],
zoomControl: false,
attributionControl: false
});
CartoDB.js raises the following error:
Uncaught TypeError: Cannot set property '_attributions' of undefined
This scenario happens in a use case in which the user is using GMaps + Leaflet at the same time. As the GMaps attribution already exists, the user decided to remove the Leaflet one.
There are a couple of problems that @iriberri discovered that I'm going to fix:
[ ] Leaflet map might have attributionControl set to false and CartoDB.js should still handle attributions correctly.
[ ] Leaflet map might have some other layers with attributions, and these shouldn't be removed.
|
2025-04-01T04:10:15.665288
| 2015-11-11T14:31:09
|
116345393
|
{
"authors": [
"alonsogarciapablo",
"rochoa",
"viddo",
"xavijam"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13733",
"repo": "CartoDB/cartodb.js",
"url": "https://github.com/CartoDB/cartodb.js/pull/791"
}
|
gharchive/pull-request
|
Encodes filters param when instantiating a dashboard
Fixes #778.
@rochoa will this work? Thanks!
@xavijam can you take a look at this please?
We are already doing encodeURIComponent in some filters: https://github.com/CartoDB/cartodb.js/blob/86d02695231f4b2cb0a6d1d8754914bff688e195/src/windshaft/filters/category.js#L108-L109
I aim for a general solution but we have to check we're not double encoding it.
Agree, we should create a general rule for this. Maybe included in the filter toJSON thing.
Why not leave the encoding to when the query string is actually needed instead of having to do this manually? e.g. using https://github.com/sindresorhus/query-string or similar
|
2025-04-01T04:10:15.667444
| 2017-11-03T07:49:48
|
270899555
|
{
"authors": [
"ethervoid",
"javitonino",
"juanignaciosl"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13734",
"repo": "CartoDB/cartodb",
"url": "https://github.com/CartoDB/cartodb/pull/13033"
}
|
gharchive/pull-request
|
Share dataset problems in both cloud and onpremise after removing groups
This fixes the issue causing CartoDB/support/issues/1057.
Answered. Take a look and move to acceptance if ok.
Deployed master to the same server and the error goes away, so it seems somewhat related to this PR. Back to dev.
@ethervoid CR or unnasign this, please ๐ค
Im on it trying to decipher the last commit xDD
Working fine. Tested group creation/deletion/rename/changing members, in conjuction with adding/removing permissions. Happy path works, and I was not able to break it in any other way. ๐
|
2025-04-01T04:10:15.678199
| 2015-07-01T13:32:30
|
92386661
|
{
"authors": [
"Cartofante",
"alonsogarciapablo",
"juanignaciosl",
"xavijam",
"ztephm"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13735",
"repo": "CartoDB/cartodb",
"url": "https://github.com/CartoDB/cartodb/pull/4251"
}
|
gharchive/pull-request
|
Possibility to save all map layers at the same time
Basically we have a problem when we need to save the state of all layers those belong to a visualization.
We used to sent as many PUTs as layers we have in the visualization. From now on, we want to just make a request with the needed info.
Fixes #4016.
@juanignaciosl
So, with the current changes the request works correctly but:
Map is not refreshed.
Layers order is not being render properly, mainly because:
https://github.com/CartoDB/cartodb/blob/4016-Possibility_to_save_all_map_layers_at_the_same_time/lib/assets/javascripts/cartodb/table/layer_panel_view.js#L356
Seems like we need to sort layers or change that part of code and get 'order' attribute.
Frontend tests were OK :+1: (details)
Frontend tests were OK :+1: (details)
Hey! What's the status of this PR? @xavijam are you planning to do some more work here? (I'm asking because of the two caveats that you mentioned in your comment). This is somehow related to this PR, so it'd be nice to QA everything at the same time.
Hi, this issue is happening still for S/B 5831712
Think you guys already got this part, but:
Expected:
drag top layer to bottom of layer stack in Map View
data layers in map re-order according to new layer order
Happening:
drag top layer to bottom of layer stack in Map View
data layers are not re-ordering
Replicated here:
Add line layer
Add point layer
Add polygon layer
Drag polygon layer to bottom
Polygon layer shows at bottom of Map View layer stack but polygons are still on top of other line and point layers:
http://cdb.io/1I3nhYG
Frontend tests were OK :+1: (details)
Frontend tests were OK :+1: (details)
I think this is now ready and working! :relieved:
@xavijam Could you PTAL? Thanks!
Looks great man :)
Frontend tests were OK :+1: (details)
Frontend tests were OK :+1: (details)
\o/ :dancer: :grapes:
Frontend tests were OK :+1: (details)
Deployed and working :+1:. Great job guys!
Yahoo! :-)
|
2025-04-01T04:10:15.679463
| 2016-09-09T10:01:27
|
175973087
|
{
"authors": [
"jgoizueta",
"rafatower"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13736",
"repo": "CartoDB/cartodb",
"url": "https://github.com/CartoDB/cartodb/pull/9723"
}
|
gharchive/pull-request
|
Fix problems with raster overview factors
Fixes #7082
@rafatower please review this
:+1: go for it, thanks for the fix
|
2025-04-01T04:10:15.711186
| 2017-07-19T10:03:49
|
243987919
|
{
"authors": [
"jaakla",
"jatorre",
"mtehver"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13737",
"repo": "CartoDB/mobile-sdk",
"url": "https://github.com/CartoDB/mobile-sdk/issues/119"
}
|
gharchive/issue
|
Enable online update of styles
Need to consider many aspects here:
online and offline cases
different data schema versions
API server-side
This is placeholder issue, for details see Feature doc
My point in here:
Our mobile SDK supports by default 4-8 styles that we maintain. Those styles live together with the server of vector tiles and are alwasy updated to work with the current schema we are providing.
If the schema changes, our published styles on the web will also be updated if they need to.
Mobile SDK applications when loading a map "voyager" for example, first retrieves the style, which includes info about the sources too, and then starts making requests to the source.
The Mobile SDK can provide a set of options to customize the view, like the list of params we said to have for our own styling of raster and on the web, and the developer have the possibility to override those parameter programatically.
If the developer decides to use an entirely different style created by him he is under risk that in the future the schema changes will break it and is its reponsability to keep it updated.
On offline. When downloading an offline package ideally the style associated with it is also downloaded, which matches the package that has been downloaded. The further customizations still apply like before.
The key issue with loading online styles is that we cannot show map at all without style, but users expect immediate maps and they do not have fast enough bandwidth to download styles before maps. Style packages can be easily quite big - current minimal is few megabytes as minimum, and can be tens of megabytes due to the bundled resources (e.g. international fonts). Mapbox has to download 50-80MB style resources before they can download single real map tile offline, and this is terrible; do not want to repeat the mistake of adding offline support as afterthought.
The biggest specific challenge in vector map styling is to have fonts (full international fontsets can go to hundreds of MB), as we cannot use device system fonts due to platform API restrictions. Now once we have some key fontset bundled with SDK, then maybe can have quick enough style loading + lazy additional custom font loading approach. Especially if we can limit at least ourselves to small number of used fonts.
Once big font/resource challenge is resolved, then core style css/xml with colors, rules etc is quite small and we can load at least our small set of styles on-demand as you suggest.
Can it be that our Mobile SDK bundless the assets so it is part of the
download of the app? At some point those MB need to be downloaded.
We dont need to have all fonts in the world avsailable, just a few that
works well... so dont see, similar to what we are saying with a few
pre-defined styles than can only be modified with parameters, I dont see
why we can not bundle a few fonts with international support that are what
they can be used straight away.
CARTO http://carto.com - Turn Location Data Into Business Outcomes
Request a DEMO now https://carto.com/request-live-demo/ to learn how we
can help you with your location data
On Wed, Jul 19, 2017 at 4:35 PM, Jaak Laineste<EMAIL_ADDRESS>wrote:
The key issue with loading online styles is that we cannot show map at all
without style, but users expect immediate maps and they do not have fast
enough bandwidth to download styles before maps. Style packages can be
easily quite big - current minimal is few megabytes as minimum, and can be
tens of megabytes due to the bundled resources (e.g. international fonts).
Mapbox has to download 50-80MB style resources before they can download
single real map tile offline, and this is terrible; do not want to repeat
the mistake of adding offline support as afterthought.
The biggest specific challenge in vector map styling is to have fonts
(full international fontsets can go to hundreds of MB), as we cannot use
device system fonts due to platform API restrictions. Now once we have some
key fontset bundled with SDK, then maybe can have quick enough style
loading + lazy additional custom font loading approach. Especially if we
can limit at least ourselves to small number of used fonts.
Once big font/resource challenge is resolved, then core style css/xml with
colors, rules etc is quite small and we can load at least our small set of
styles on-demand as you suggest.
โ
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/CartoDB/mobile-sdk/issues/119#issuecomment-316508847,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAHzOxLMMVKnP8cC6lB06W1P-Y16FqNcks5sPmh-gaJpZM4Ocgd5
.
Yes, bundling as much fonts as we can is what I meant also. Full international set is big, so far have got smallest package (one font covering all countries with official languages) to about 5MB. Compared to SDK binary itself (about 4MB) it would more than double SDK, and some developers are very sensitive about it, as they want to keep whole app install in Android below 10MB limit.
Another challenge is that inside SDK APIs and classes for data and styles are independent by design, adding coordinated versioning solution there involves changing many pieces, see drawing in the feature doc (link is above). It can, and should be done, just needs some detailed design.
@mtehver - updated feature doc. Next steps:
[ ] Prepare script to generate "files" section with file sizes and md5-s @mtehver
[ ] Update metadata jsons: tile.json, packages.json in server (S3) - I can do
[ ] Create style service - it is static files, could be S3 files like with the tile/packages.json - I can do
[ ] Update SDK side to load new files and update existing styles as in feature doc. @mtehver
[ ] Create small doc guideline for "content management" how to update styles for mobile and how to deploy. Deploy is probably saving to GIT, then run script to update md5 and then upload to S3
The script to generate 'files' section in style.json is under mobile-styles repository (tools folder). SDK in develop branch is now updated to support online style updates according to the design doc.
Server-side http://api.nutiteq.com/styles/carto.streets/v1/voyager.json works now, files are in Azure Storage.
Note that in this implementation custom (new) styles cannot be updated, only SDK built-in styles (3 styles) can. So to add new styles also SDK needs small code change, app will need SDK update, and then can add new styles. This should be ok, as the apps could not use new styles without app code update anyway. This can be added when there is real life requirement for it.
|
2025-04-01T04:10:15.763714
| 2020-11-04T15:45:37
|
736219461
|
{
"authors": [
"CastagnaIT",
"neogeo71"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13738",
"repo": "CastagnaIT/plugin.video.netflix",
"url": "https://github.com/CastagnaIT/plugin.video.netflix/issues/928"
}
|
gharchive/issue
|
firestick that plays 1080p just fine,new firetvcube 2019 will not play higher than 540p
i have a firestick that plays 1080p just fine,new firetvcube 2019 will not play higher than 540p.
I tried inputting my esn from netflix and same result
I installed the material terminal and ran getprop and see the output but there is no way to email from firetv. I cannot find a working email app. any ideas how I can save this terminal output to get it to you to review?
i tried getprop >info.txt but it says it is read only and cannot write the file.
please advise? I really would like to get your plugin working. I think it is a cube specific issue as I found one other user on reddit with same issues.
Thank you in advance for your time and attention to this matter.
I need you provide specified info otherwise I can not help
mandatory is a Debug Log follow all the steps explained in the Readme
but take the log only after you have done Reset ESN from expert setting
Follow step 1 and 3 of Video Stream only at 540P (SD) on Android device:
https://github.com/CastagnaIT/plugin.video.netflix/wiki/FAQ-(Audio%2C-Video%2C-Subtitle%2C-Other)#video-stream-only-at-540p-sd-on-android-device
and so attach/report all required info
This is my issue:
Install "Material Terminal" app or similar terminal app
Open the app, in the black screen write getprop so a long list appears
Open the options then select "Send to e-mail". (you can send the info to yourself and attach everything as a text file or if you prefer send directly to this GitHub e-mail)
I cannot do step 3 on the FireTV cube. How do I get this output off the FireTV to send to you?
alternative are:
try to connect a usb/bluetooth mouse
install TeamViewer to firecube and your phone, then control your firecube from the phone, in order to open the "material terminal" options
not easy alternative, use ADB via IP to control the shell from a computer, explain how to here require lot time, if you want try this way look from the web
closed due to inactivity
|
2025-04-01T04:10:15.785342
| 2020-04-20T21:01:53
|
603532891
|
{
"authors": [
"JiDW",
"moo-man"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13740",
"repo": "CatoThe1stElder/WFRP-4th-Edition-FoundryVTT",
"url": "https://github.com/CatoThe1stElder/WFRP-4th-Edition-FoundryVTT/issues/384"
}
|
gharchive/issue
|
Add a system for spending "Fortune" points
There is currently two things (out of 3) that could be handled by the system:
Rerolling a test
Adding +1 SL to a test
I suggest to add either:
a contextual menu to a roll with the following items available for the players who rolled =>
Reroll (-1 Fortune)
Add +1 SL (-1 Fortune)
or
a private card chat after a roll with the following elements =>
Two buttons for reroll + adding SL
The remaining fortune points
Either way, another chat card is send to everybody to let them know that a fortune point has been used and how.
For a reroll, we also send a new roll (same target/modifiers)
For a +1SL, I suggest we just edit the previous roll to limit spams in the chat.
I'm not sure about the contextual vs card.
I'm not a fan of contextual menu because it is hidden from the players and it requires more training time. The GM has to know about it and tell his players
But another chat card, even if private, could be a bit much considering we already have a lot of chat card sent during a fight.
I think I still prefer the chat card tho, any thoughts?
I dislike the extra chat card, as you said, it would clutter up the chat menu and be, to me, incredibly annoying to see it for every single roll you do.
I am definitely in support of the contextual menu, I'm not sure about your reservations, the 'training' needed is very minimal. It may be hidden, but there can be ways to make it more apparent, such as an icon+tooltip to tell the user that they can right click this card.
Oh yeah, an icon is a very good idea, best of both worlds!
Well, it could be only me but the first time I discovered your system, the only thing that lost me was the "Apply damage" contextual menu. I figured it out thanks to a tutorial video but everything else was so simple that I didn't think of that in the first place while discovering the system without reading anything about it
|
2025-04-01T04:10:15.802363
| 2021-12-04T01:26:30
|
1071084128
|
{
"authors": [
"Thiago-Reis-Porto",
"zahrahadavand"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13741",
"repo": "CcgAlberta/pygeostat",
"url": "https://github.com/CcgAlberta/pygeostat/issues/31"
}
|
gharchive/issue
|
infergriddef - 2D data Problem
Hi, I think I have found a problem with the "infergriddef" method, if an Integer is passed to the "nblk" parameter. With an integer, the nz will never be none and "twod" will be False, raising an exception if it's an 2D data.
Also, the same happens with float or if the z value is passed as 1 on the blksize. So the documentation example doesn't work.
In docstring of "pygeostat/data.py/infergriddef", this is mentioned that:
"If data is 2-D, zsiz or nz must be provided as None. Otherwise it raise exception:"
The examples have been corrected in the documentation: (the corrected examples in the docstring will be updated to the next release of pygeostat)
df2d = gs.ExampleData("point2d_ind")
df2d.infergriddef(nblk = [60, 60, None])
df2d.infergriddef(blksize = [50,60,None])
|
2025-04-01T04:10:15.824459
| 2021-02-09T18:11:01
|
804791037
|
{
"authors": [
"XiaoningDing",
"sonyafenge"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13743",
"repo": "CentaurusInfra/arktos",
"url": "https://github.com/CentaurusInfra/arktos/pull/981"
}
|
gharchive/pull-request
|
Promote admission webhook API to v1
What type of PR is this?
/kind feature
What this PR does / why we need it:
This PR is part of kube-openapi upgrade: #923
all information cherry-picked from community kubernetes PR: https://github.com/kubernetes/kubernetes/pull/79549
Which issue(s) this PR fixes:
Fixes #923
Special notes for your reviewer:
Does this PR introduce a user-facing change?:
/lgtm
/approve
|
2025-04-01T04:10:15.826090
| 2022-05-18T23:19:01
|
1240699409
|
{
"authors": [
"davidzhu22",
"h-w-chen"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13744",
"repo": "CentaurusInfra/fornax-serverless",
"url": "https://github.com/CentaurusInfra/fornax-serverless/issues/38"
}
|
gharchive/issue
|
[fornax core] etcd client for namespace and application
package of fornax core etcd client, able to Create/Get/Update/Delete/List/Watch following data types:
Namespace
Application
( endpoint type will be needed later )
typical use scenario
client := foo.NewEtcdClient(...)
client.CreateNamespace(namespace)
client.CreateApplication(app)
...
apps := client.ListApplications(namespace) // of specified namespace
chApps := client.Watch(nil) // of all namespaces
app<-chApps
I will follow the rule to do it
|
2025-04-01T04:10:15.832682
| 2014-10-16T20:19:01
|
46030654
|
{
"authors": [
"csheldonhess",
"nanchen47"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13745",
"repo": "CenterForOpenScience/osf.io",
"url": "https://github.com/CenterForOpenScience/osf.io/issues/1127"
}
|
gharchive/issue
|
Emoji are allowed in titles but do not display in all browsers
I named my project "This project has a wiki, ๐" and my wiki "This is A CoOoOoOol WiKi ๐"
I entered the information in Chrome, so the emoji came up as boxes, rather than what I picked. I expected they would be boxes in all of the other browsers. However, had I entered them in Safari or Firefox and seen them show up, I would have expected them to show up in Chrome and Opera, too.
I confirmed that this is the case on Staging and Production.
In Safari and Firefox (on Mac) they both show up fine:
In Chrome and Opera (on Mac) they do not show up:
I confirmed that this is the case with emoji, but not with proper Unicode characters, which show up just fine, no matter which of the four browsers I used:
After a quick Google search, it seems that Chrome does not natively support emoji. Perhaps we should not allow them in the OSF, to keep things consistent.
Will be considered in later enhancement. Thanks.
|
2025-04-01T04:10:15.838245
| 2015-01-19T19:28:36
|
54802868
|
{
"authors": [
"caileyfitz",
"nanchen47"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13746",
"repo": "CenterForOpenScience/osf.io",
"url": "https://github.com/CenterForOpenScience/osf.io/issues/1524"
}
|
gharchive/issue
|
[staging & production] Consistency in add-on deletion modal for user settings page
Steps
Enable multiple add-ons for two or more projects.
Go to one of the projects where add-ons are enabled.
In the Settings tab, go to the Select Add-ons section and deselect two or more add-ons and press Submit.
Notice what happens after the Submit button is clicked.
Now, go to the user settings page and navigate to Configure Add-ons.
Deselect two or more add-ons and press Submit.
Notice what happens after the Submit button is clicked.
Expected
I expect that the behavior would be the same when deselecting add-ons from project settings vs. the user settings page, ie: in terms of modals, success messages, etc.
Actual
Deselecting add-ons from these two places does not yield consistent behavior. When deselecting from the project settings page, pressing Submit immediately removes the add-ons and then a success message appears. When deselecting from the user settings page, pressing Submit prompts a confirmation modal:
listing the add-ons to remove.
I think the behavior here should be consistent and that a confirmation modal should be added when deselecting add-ons via the project settings page, just as a precaution against accidental add-on removal.
Additionally, I think that users could benefit from more information in the confirmation modal on the user settings page. The current modal does not adequately point to the potentially wide-spread reach of removing add-ons. For example, imagine that a professor in a lab authorizes his own Dataverse account for a project that his graduate students are working on. The professor doesn't work on the project in the OSF and is in fact not even an admin on the project -- the grad students are doing all of the uploading/downloading/editing on the site. Then, maybe the prof wants to clean up and deselects Dataverse, completely forgetting that it is his account which is linked for the project. There is no indication that deselecting the add-ons will remove them from every project to which his accounts are linked. Even just a sentence saying something like "Removing these add-ons will remove them from every project to which your accounts are linked, regardless of your permissions on the project -- are you sure you want to remove these add-ons?" could prevent users from accidentally removing add-ons when they didn't actually mean to.
same issue as https://github.com/CenterForOpenScience/osf.io/issues/1621
|
2025-04-01T04:10:15.853911
| 2018-07-13T10:33:14
|
340968211
|
{
"authors": [
"AljoLe",
"felliott"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13747",
"repo": "CenterForOpenScience/osf.io",
"url": "https://github.com/CenterForOpenScience/osf.io/issues/8543"
}
|
gharchive/issue
|
git repositories can't be cloned inside *_requirements containers
Hey there,
I'm trying to install OSF on a local machine, in order to try things out.
When I run docker-compose up requirements mfr_requirements wb_requirements git throws the following warnings:
Inside the requirements_1 container:
requirements_1 | fatal: unable to access 'https://github.com/CenterForOpenScience/dataverse-client-python.git/': gnutls_handshake() failed: The TLS connection was non-properly terminated.
requirements_1 | Command "git clone -q https://github.com/CenterForOpenScience/dataverse-client-python.git /code/src/dataverse" failed with error code 128 in None
Inside the mfr_requirements_1 container:
mfr_requirements_1 | fatal: unable to access 'https://github.com/CenterForOpenScience/waterbutler.git/': gnutls_handshake() failed: The TLS connection was non-properly terminated.
mfr_requirements_1 | Command "git clone -q https://github.com/CenterForOpenScience/waterbutler.git /tmp/pip-install-3xp7cbxp/waterbutler" failed with error code 128 in None
and inside the wb_requirements_1 container:
wb_requirements_1 | fatal: unable to access 'https://github.com/felliott/boto.git/': gnutls_handshake() failed: The TLS connection was non-properly terminated.
wb_requirements_1 | Command "git clone -q https://github.com/felliott/boto.git /tmp/pip-install-r1w23mjg/boto" failed with error code 128 in None
It seems, that git can't reach the repos from inside of the containers.
Changing line 4 in ./addons/dataverse/requirements.txt from
-e git+https://github.com/CenterForOpenScience/dataverse-client-python.git@5f20ad7aba11b0ce0bee6e791a91266e7ce4d76a#egg=dataverse
to
-e git+http://github.com/CenterForOpenScience/dataverse-client-python.git@5f20ad7aba11b0ce0bee6e791a91266e7ce4d76a#egg=dataverse
helps to solve the problem for the requirements_1-container, but I can't find the files for the other git repositories, in order to change them from httpsto http.
The whole log file is attached here:
docker-compose-requirements-log.txt
I hope you can help me with that, but maybe it's just a docker config thing.
Thanks in advance and best regards from Hamburg,
AljoLe
Hey @AljoLe,
That looks like something is running an old version of tls. Are your images up-to-date? A couple of things to try:
Try running git clone https://github.com/felliott/boto.git from your command line to see if it's something on the network at large, or something specific to the container.
Start a bash shell in the wb container: docker exec -it "osf_wb_1" /bin/bash. cd to the tmp directory and run git clone https://github.com/felliott/boto.git. Does that work?
If neither works, you might be behind a proxy that doesn't speak tls v1.2. If only the first works, then it might be something in the container setup. Let's try those first and we'll figure out where to go from there.
Cheers,
Fitz
Hey @felliott,
thanks for your response.
Here is a list of my images:
[al@localhost osf.io]$ docker images
REPOSITORY TAG IMAGE ID CREATED SIZE
quay.io/centerforopenscience/osf-web develop-local 70fddbcdd397 2 days ago 3.64GB
quay.io/centerforopenscience/mfr develop 79d20fae2403 2 days ago 2.19GB
quay.io/centerforopenscience/osf develop 6941deb51ea6 3 days ago 1.44GB
quay.io/centerforopenscience/osf-preprints develop-local d7a7f8a535b2 4 days ago 2.3GB
quay.io/centerforopenscience/wb develop 3cf3f2ae2333 5 days ago 528MB
mongo 3.4 f6eb71f3b2e9 5 days ago 361MB
rabbitmq 3-management 68bdffcb99c0 9 days ago 149MB
quay.io/centerforopenscience/osf-registries develop-local 96fae2a699d8 10 days ago 2.42GB
elasticsearch 2 bc337c8e4f39 13 days ago 479MB
postgres 9.6 cfb9cf0ded21 13 days ago 235MB
quay.io/centerforopenscience/sharejs develop 56d5a1af2eea 5 months ago 673MB
quay.io/centerforopenscience/fakecas master<PHONE_NUMBER>e3 13 months ago 767MB
and of my containers:
[al@localhost osf.io]$ docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
648d628d9c7c quay.io/centerforopenscience/osf-preprints:develop-local "/bin/bash -c 'yarn โฆ" 2 hours ago Exited (137) About an hour ago osfio_preprints_1
44bf33d30aa2 quay.io/centerforopenscience/osf-web:develop-local "/bin/bash -c 'yarn โฆ" 2 hours ago Exited (137) About an hour ago osfio_ember_osf_web_1
55ac3107bd09 quay.io/centerforopenscience/sharejs:develop "gosu node npm start" 2 hours ago Exited (137) About an hour ago osfio_sharejs_1
b10850fd7a74 quay.io/centerforopenscience/osf-registries:develop-local "/bin/bash -c 'yarn โฆ" 2 hours ago Exited (137) About an hour ago osfio_registries_1
a6564d67489a quay.io/centerforopenscience/osf:develop "invoke admin.assetsโฆ" 2 hours ago Exited (137) About an hour ago osfio_admin_assets_1
91c5977263e6 quay.io/centerforopenscience/wb:develop "invoke server" 2 hours ago Up About an hour <IP_ADDRESS>:7777->7777/tcp osfio_wb_1
a33c35717d32 quay.io/centerforopenscience/osf:develop "invoke assets -dw" 2 hours ago Exited (137) About an hour ago osfio_assets_1
438bac8a2374 quay.io/centerforopenscience/osf:develop "invoke server -h 0.โฆ" 2 hours ago Exited (137) About an hour ago osfio_web_1
5ebeb663aedc quay.io/centerforopenscience/fakecas:master "fakecas -host=0.0.0โฆ" 2 hours ago Exited (0) About an hour ago osfio_fakecas_1
a700bb3cc4b6 quay.io/centerforopenscience/mfr:develop "invoke server" 2 hours ago Exited (0) About an hour ago osfio_mfr_1
0a1a19039204 quay.io/centerforopenscience/osf:develop "invoke celery_worker" 2 hours ago Restarting (1) About an hour ago osfio_worker_1
e6af80f3b135 quay.io/centerforopenscience/osf:develop "invoke adminserver โฆ" 2 hours ago Restarting (1) About an hour ago osfio_admin_1
f3608fbe4475 quay.io/centerforopenscience/osf:develop "invoke apiserver -hโฆ" 2 hours ago Exited (137) About an hour ago osfio_api_1
1bfec4dfe3f4 quay.io/centerforopenscience/osf:develop "-i -t /bin/bash" 4 hours ago Created pedantic_pasteur
f01a4f203baa quay.io/centerforopenscience/wb:develop "/bin/bash -c 'invokโฆ" 4 hours ago Exited (1) 3 hours ago osfio_wb_requirements_1
e2a664433842 quay.io/centerforopenscience/mfr:develop "/bin/bash -c 'invokโฆ" 4 hours ago Exited (137) 4 hours ago osfio_mfr_requirements_1
937d0c7fdc65 quay.io/centerforopenscience/osf:develop "/bin/bash -c 'invokโฆ" 4 hours ago Exited (1) 3 hours ago osfio_requirements_1
Running git clone https://github.com/felliott/boto.gitfrom the command line works flawlessly.
Starting a bash shell in the wb container works, but I have some problems, that might be relevant. After running docker-compose up -d wb; docker exec -it "osf_wb_1" /bin/bash, the CLI tells me:
[al@localhost osf.io]$ docker-compose up -d wb; docker exec -it "osf_wb_1" /bin/bash
osfio_wb_1 is up-to-date
Error: No such container: osf_wb_1
so, there is no container called osf_wb_1, but there is a container called osfio_wb_1.
Running
[al@localhost osf.io]$ docker-compose up -d wb; docker exec -it "osfio_wb_1" /bin/bash
osfio_wb_1 is up-to-date
root@91c5977263e6:/code#
brings me to the container. In there I can change to the tmpdirectory and try to start the git clonecommand:
root@91c5977263e6:/code# cd ../tmp/
root@91c5977263e6:/tmp# git clone https://github.com/felliott/boto.git
Cloning into 'boto'...
fatal: unable to access 'https://github.com/felliott/boto.git/': gnutls_handshake() failed: The TLS connection was non-properly terminated.
root@91c5977263e6:/tmp#
So all in all I can say, that cloning a git repository works totally fine in my ~ directory, but inside the docker container, git can clone anything, even when I use httpinstead of 'https`
I hope, I delivered some usefull information.
Thanks and best regards,
Aljoscha
Just to clarify some things,
I cloned the osf.io git repository with git clone https://github.com/CenterForOpenScience/osf.io.git
and worked through README-docker-compose.md.
To come back to your questions, here is the output of the commands you asked for:
[al@localhost osf.io]$ docker-compose up -d wb; docker exec -it "osfio_wb_1" /bin/bash
Creating osfio_wb_1 ... done
root@e6dfd7c3782b:/code# curl -v https://github.com
bash: curl: command not found
root@e6dfd7c3782b:/code# apt-cache show openssl
Package: openssl
Status: install ok installed
Priority: optional
Section: utils
Installed-Size: 1020
Maintainer: Debian OpenSSL Team<EMAIL_ADDRESS>Architecture: amd64
Version: 1.0.1t-1+deb8u8
Depends: libc6 (>= 2.15), libssl1.0.0 (>= 1.0.1k-3+deb8u3)
Suggests: ca-certificates
Conffiles:
/etc/ssl/openssl.cnf 7df26c55291b33344dc15e3935dabaf3
Description: Secure Sockets Layer toolkit - cryptographic utility
This package is part of the OpenSSL project's implementation of the SSL
and TLS cryptographic protocols for secure communication over the
Internet.
.
It contains the general-purpose command line binary /usr/bin/openssl,
useful for cryptographic operations such as:
* creating RSA, DH, and DSA key parameters;
* creating X.509 certificates, CSRs, and CRLs;
* calculating message digests;
* encrypting and decrypting with ciphers;
* testing SSL/TLS clients and servers;
* handling S/MIME signed or encrypted mail.
Description-md5: 9b6de2bb6e1d9016aeb0f00bcf6617bd
root@e6dfd7c3782b:/code# apt-cache search tls
libssl1.0.0 - Secure Sockets Layer toolkit - shared libraries
libcurl3-gnutls - easy-to-use client-side URL transfer library (GnuTLS flavour)
libssl-dev - Secure Sockets Layer toolkit - development files
libksba8 - X.509 and CMS support library
libdevmapper1.02.1 - Linux Kernel Device Mapper userspace library
libevent-openssl-2.0-5 - Asynchronous event notification library (openssl)
libgnutls-deb0-28 - GNU TLS library - main runtime library
openssl - Secure Sockets Layer toolkit - cryptographic utility
libssl-doc - Secure Sockets Layer toolkit - development documentation
root@e6dfd7c3782b:/code#
root@e6dfd7c3782b:/code# curl -v https://github.com
bash: curl: command not found
Apologies, I forgot we uninstall curl in the Dockerfile. Try this:
$ apt-get update
$ apt-get install -y curl
$ curl -v https://github.com
Cheers,
@felliott
No problem,
the output from
$ curl -v https://github.com
is quite long, so you will find it in this file:
log.curl.github.txt
Hope that helps,
good night
@AljoLe
|
2025-04-01T04:10:15.856747
| 2016-06-22T19:37:09
|
161767542
|
{
"authors": [
"brianjgeiger"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13748",
"repo": "CenterForOpenScience/osf.io",
"url": "https://github.com/CenterForOpenScience/osf.io/pull/5880"
}
|
gharchive/pull-request
|
Hotfix/Remove devModeControls when not in dev mode
Purpose
Removes the devModeControls div when not in master so the previous fix will apply
Changes
Removed the else branch on the mako template that added an empty div. Tested by manually removing all divs to ensure that nothing was loaded and removed the file to make sure file didn't 404 if the div didn't exist.
Side effects
No, should be good.
Ticket
https://openscience.atlassian.net/browse/OSF-6535
Cancelled tests
|
2025-04-01T04:10:15.857913
| 2015-07-07T16:05:44
|
93566163
|
{
"authors": [
"alexwelch",
"argvader",
"dharmamike"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13749",
"repo": "CenturyLinkLabs/panamax-ui",
"url": "https://github.com/CenturyLinkLabs/panamax-ui/pull/539"
}
|
gharchive/pull-request
|
cleaning up modal dialog fixtures during jasmine test runs
Adds a global afterEach to the jasmine spec_helper.js to clean up leftover DOM elements added to the document body during modal dialog creation. These elements were leaking into other test fixtures and causing failures.
:+1:
:+1:
|
2025-04-01T04:10:15.906285
| 2022-06-21T21:20:44
|
1279112475
|
{
"authors": [
"ggetz",
"techmavengeospatial"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13750",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/issues/10472"
}
|
gharchive/issue
|
Load LERC Elevation-Terrain Tiles from tile server (XYZ) not just from ImageServer - ArcGISTiledElevationTerrainProvider
If ESRI has their data available like this
http://elevation3d.arcgis.com/arcgis/rest/services/WorldElevation3D/Terrain3D/ImageServer/tile/{z}/{y}/{x}.lerc2
why can't I serve data to Cesium like this?
http://someip:port/folder/tile/{z}/{y}/{x}.lerc2
https://github.com/CesiumGS/cesium/blob/3601fc73247ab3526c4f18be38ccc8c1860a541b/Source/Core/ArcGISTiledElevationTerrainProvider.js#L63
https://cesium.com/learn/cesiumjs/ref-doc/ArcGISTiledElevationTerrainProvider.html
https://github.com/CesiumGS/cesium/blob/main/Apps/Sandcastle/gallery/ArcGIS Tiled Elevation Terrain.html
https://cesium.com/learn/cesiumjs/ref-doc/Resource.html
I've produced my own LERC tiles and I need to have an on-premises/within firewall and mobile solution
convert Gridded GeoTIFF Tiles to LERC with this C++ Code
GitHub - DaYeSquad/tif2lerc-cpp: Convert tiff to esri lerc format
convert Gridded GeoTIFF Tiles into LERC with Pyton code from ESRI
GitHub - Esri/lerc: Limited Error Raster Compression
Reference also Terriajs we use Terriajs in other projects https://github.com/TerriaJS/terriajs/discussions/4890
Hi @techmavengeospatial,
From those threads you linked, there seems to be a problem when Cesium requests the first tile at level=0, Y=0, X=0 and that tile is not available on the server, so the request fails. Due to the failure, Cesium stops requesting tiles. Here's a sandcastle showing this behavior.
Do you have an example with your own data, or can you confirm that's the case with your own data?
|
2025-04-01T04:10:15.911085
| 2022-02-03T18:45:23
|
1123435058
|
{
"authors": [
"cesium-concierge",
"j9liu",
"ptrgags"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13751",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/pull/10071"
}
|
gharchive/pull-request
|
Add support for debugColorizeTiles in ModelExperimental
Fixes #10061.
This PR adds support for debugColorizeTiles in ModelExperimental and re-enables its unit tests. Here's a sandcastle with the same glTF model used in the unit test; previously, no colors were applied when Display > Colorize as selected in the insepctor.
Thanks for the pull request @j9liu!
:heavy_check_mark: Signed CLA found.
:grey_question: CHANGES.md was not updated.
If this change updates the public API in any way, please add a bullet point to CHANGES.md.
Reviewers, don't forget to make sure that:
[ ] Cesium Viewer works.
[ ] Works in 2D/CV.
Changes look good. Now that Travis is passing, I'm merging this PR. Thanks @j9liu!
|
2025-04-01T04:10:15.920839
| 2023-02-09T05:21:21
|
1577261367
|
{
"authors": [
"cesium-concierge",
"ggetz",
"jjhembd",
"onsummer"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13752",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/pull/11080"
}
|
gharchive/pull-request
|
[WIP] Implement typescript types
Make types more precision.
TODO:
[x] Support TypeScript import() type in JSDoc comments
[ ] Add correct types definitions for unclear API such as fabric JSON Object in Material, etc.
Thanks for the pull request @onsummer!
:heavy_check_mark: Signed CLA found.
:grey_question: CHANGES.md was not updated.
If this change updates the public API in any way, please add a bullet point to CHANGES.md.
:grey_question: Unit tests were not updated.
Make sure you've updated tests to reflect your changes, added tests for any new code, and ran the code coverage tool.
Reviewers, don't forget to make sure that:
[ ] Cesium Viewer works.
[ ] Works in 2D/CV.
Hi @onsummer, thanks for the effort to fix up some typing issues!
I would appreciate if you could explain in a bit more detail here what changes you are making to the tooling, and why.
@ggetz OK, poor for my English, I'll try to explain by edit above as concisely as possible
@ggetz PTAL. I found that adding additional types (eg. callback functions' type) may require a new PR, what do you think?
@onsummer, yes. I think it would be helpful for review to separate the tooling changes from the type updates. Thanks!
Thanks @onsummer!
One thing I'm still not clear on is why jsdoc-tsimport-plugin is needed. I don't see where we are using the import syntax. Could you please explain?
@ggetz Sorry, forget to remove that. This dependency should appearance in next PR that will improve types by using import() syntax.
@jjhembd Thank you for your review. I fix some issues that you mentioned above as possible as I can found, PTAL
Hi @onsummer, thanks for those changes!
I pointed out a few small remaining things in ModelComponents and ScreenSpaceCameraController. Once those are done, this will be ready to merge.
Hi @onsummer, thanks again for this contribution. I learned something about how to write better JSDoc just by looking at your edits.
We would like to make sure future code changes follow the same pattern. We have an open issue to automate JSDoc checks in #10815. Once this PR is merged, would you be interested in contributing to that issue?
Hi @onsummer, thanks again for this contribution. I learned something about how to write better JSDoc just by looking at your edits.
We would like to make sure future code changes follow the same pattern. We have an open issue to automate JSDoc checks in #10815. Once this PR is merged, would you be interested in contributing to that issue?
That seems a pretty challenge to me, and I am interesting in contributing it.
Hi @onsummer, thanks for those changes! I pointed out a few small remaining things in ModelComponents and ScreenSpaceCameraController. Once those are done, this will be ready to merge.
:) Maybe this time will be OK to merge
Looks good, thanks a lot @onsummer! Merging now
|
2025-04-01T04:10:15.925463
| 2024-04-17T14:43:17
|
2248476315
|
{
"authors": [
"ggetz",
"jjspace"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13753",
"repo": "CesiumGS/cesium",
"url": "https://github.com/CesiumGS/cesium/pull/11939"
}
|
gharchive/pull-request
|
Don't copy .d.ts files with other assets in build
Description
The new Check.d.ts file was accidentally included in the build into the Source directory. This prevents that
Issue number and link
Fixes https://github.com/CesiumGS/cesium/issues/11918#event-12358663001
Testing plan
Clean your repo locally
run npm run make-zip
make sure Source/Core/Check.d.ts is not created
Author checklist
[ ] I have submitted a Contributor License Agreement
[ ] I have added my name to CONTRIBUTORS.md
[ ] I have updated CHANGES.md with a short summary of my change
[ ] I have added or updated unit tests to ensure consistent code coverage
[ ] I have update the inline documentation, and included code examples where relevant
[ ] I have performed a self-review of my code
Looks good! Thanks @jjspace!
|
2025-04-01T04:10:16.030271
| 2020-08-13T13:10:29
|
678425000
|
{
"authors": [
"mpetrunic"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13754",
"repo": "ChainSafe/lodestar",
"url": "https://github.com/ChainSafe/lodestar/pull/1401"
}
|
gharchive/pull-request
|
Faster attestation validation and processing
Since this is quite extensive change, I would like some sanity review before I start refactoring tests for this new logic.
Some noticable improvements:
introduce checkpoint state cache (a lot of attestations has same target checkpoint so we can reuse checkpoint states instead processing blocks all over again)
introduce seen attestation lru cache (for "first seen attestation for validator" - prevent database lookups)
reduce duplicate validation across board
attestation processor has both pending block and slot pool where attestation wait until they are ready
@ChainSafe/eth2-0 Should be ready now!
There is no more attestation validation errors in e2e tests so we reach finalization faster.
Unfortunately, I had to rollback changes to AggregateAndProof as they've caused e2e errors. I will tackle that in separate PR along with unit tests for it.
|
2025-04-01T04:10:16.033140
| 2021-06-15T13:30:29
|
921401098
|
{
"authors": [
"dapplion",
"mpetrunic"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13755",
"repo": "ChainSafe/lodestar",
"url": "https://github.com/ChainSafe/lodestar/pull/2705"
}
|
gharchive/pull-request
|
Fix nightly release
Motivation
Nightly release is broken.
Also the version is so long that NPM breaks, see https://github.com/ChainSafe/lodestar/issues/2703
Description
Use --no-git-reset flag to access version artifacts in next step
Use shorter commit sha
@wemeetagain @mpetrunic is the prepublishOnly script necessary? It causes an unnecessary double build only in some packages.
@wemeetagain @mpetrunic is the prepublishOnly script necessary? It causes an unnecessary double build only in some packages.
Probably not but you never know what some package needs.
|
2025-04-01T04:10:16.049251
| 2020-07-20T11:55:57
|
661734319
|
{
"authors": [
"Yougotcheezed",
"kabadesu",
"vini2003"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13756",
"repo": "Chainmail-Studios/Astromine",
"url": "https://github.com/Chainmail-Studios/Astromine/issues/163"
}
|
gharchive/issue
|
No bronze
No current recipe for bronze. Cant build circuits
Fixed.
This still seems to be a problem in astromine-1.5.2+fabric-1.16.1.jar. If this was fixed for 1.16.2 would it be possible to back port it for 1.16.1?
|
2025-04-01T04:10:16.055343
| 2017-12-21T15:59:08
|
283926236
|
{
"authors": [
"Chalarangelo",
"Y-Less",
"skatcat31",
"tchaffee"
],
"license": "CC-BY-4.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13757",
"repo": "Chalarangelo/30-seconds-of-code",
"url": "https://github.com/Chalarangelo/30-seconds-of-code/issues/294"
}
|
gharchive/issue
|
[BUG]validateEmail is wrong
[BUG] validateEmail does not follow the spec.
Expected Snippet Behavior
https://haacked.com/archive/2007/08/21/i-knew-how-to-validate-an-email-address-until-i.aspx/
Current Snippet Behavior
"me<EMAIL_ADDRESS>- apparently invalid, but should be valid (quoted prefixes).
<EMAIL_ADDRESS>- apparently valid, but should be invalid (not an IP).
<EMAIL_ADDRESS>- apparently invalid, but should be valid (any internationalised domain).
Possible Solution
Just don't try! The ONLY correct way to validate an e-mail is to send a test e-mail. I've seen page-long regexes designed to correctly verify all the parts, and that STILL doesn't guarantee that the result actually exists.
We could also possibly add a disclaimer that this is a basic check and not production-ready if we don't want to remove. Opinions/suggestions/fixes?
We could also possibly add a disclaimer that this is a basic check and not production-ready
It's not really a basic check. It's a broken check. Better to just remove it. Too much time has already been lost by developers trying to validate email addresses and it creates a bad user experience when you try to enter a valid working email address into a website only to have it rejected by the javascript as invalid.
The only real way to "validate" and email has been stated: See if you get a response from any server. In my work we decided to do a simple check to just make sure it "mostly" conformed: check for an @ symbol that is not at the start or end of the string. The RFC is huge. IPs are valid, unicode is valid, blah blah blah.
Attempting to do something like this has been seen as a fools folly in EVERY industry, just like validating a mailing address, for the exact same reason.
@Y-Less isn't
<EMAIL_ADDRESS>
invalid because the RFC states IP direct addresses need to be surrounded by brackets( [ & ] )?
I'm not sure on that - I was just pointing out that 999 (and 270) is not a valid IP component, but is still accepted as-is.
@skatcat31 Your argument and explanation is compelling enough to convince me (after reading a bit about mail validation and some opinions around here). I, too, vote to remove the snippet.
|
2025-04-01T04:10:16.086027
| 2023-01-19T04:06:39
|
1548481737
|
{
"authors": [
"Jaben",
"meghasinghal1"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13758",
"repo": "ChangemakerStudios/serilog-sinks-mongodb",
"url": "https://github.com/ChangemakerStudios/serilog-sinks-mongodb/issues/81"
}
|
gharchive/issue
|
How to set expiry on MongoDB document - Can someone please share working example of it?
Hi Team,
Need a quick help please on log document which is getting created in MongoDB.
How to set expiry on MongoDB document so it can be deleted automatically? I read that, it was rolled out part of 5.x version, Could someone please share a working example of it?
Thanks
var log = new Serilog.LoggerConfiguration().WriteTo.MongoDBBson(cfg => {
cfg.SetCollectionName("log");
// set the TTL on the log entry to 5 minutes expiration
cfg.SetExpireTTL(TimeSpan.FromMinutes(5));
}).MinimumLevel.Verbose().CreateLogger();
Closing as I provided an example.
|
2025-04-01T04:10:16.163340
| 2017-09-30T04:07:42
|
261816210
|
{
"authors": [
"NobodyNada",
"SmokeDetector"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13759",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/1109"
}
|
gharchive/pull-request
|
quartata: Watch gihosoft.com
quartata requests the watch of the watch_keyword gihosoft.com. See the Metasmoke search here
!!/approve
|
2025-04-01T04:10:16.166603
| 2024-09-25T07:32:10
|
2547191101
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13760",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/13493"
}
|
gharchive/pull-request
|
Cow: Watch webodoctor.in
Cow requests the watch of the watch_keyword webodoctor\.in. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
webodoctor\.in has been seen in 1 true positive, 0 false positives, and 0 NAAs.
Approved by Mast in Charcoal HQ
|
2025-04-01T04:10:16.169622
| 2020-05-25T07:50:55
|
624115056
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13761",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/3880"
}
|
gharchive/pull-request
|
Kulfy: Watch talkwidtech.com
Kulfy requests the watch of the watch_keyword talkwidtech\.com. See the MS search here and the Stack Exchange search here.
talkwidtech\.com has been seen in 0 true positives, 0 false positives, and 0 NAAs.
Approved by tripleee in Charcoal HQ
|
2025-04-01T04:10:16.172747
| 2020-09-14T13:41:04
|
701114994
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13762",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/4801"
}
|
gharchive/pull-request
|
Mast: Watch payop.com
Mast requests the watch of the watch_keyword payop\.com. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
payop\.com has been seen in 1 true positive, 0 false positives, and 0 NAAs.
Approved by Daniil in Charcoal HQ
|
2025-04-01T04:10:16.186022
| 2017-05-04T05:14:42
|
226179510
|
{
"authors": [
"angussidney",
"codecov-io",
"teward",
"tripleee"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13763",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/695"
}
|
gharchive/pull-request
|
gitmanager.py: add a critical section with a lock
This way, only one blacklisting request can be in progress at any time. Further requests will simply block on attempting to acquire the lock, and thus wait nicely in queue.
Fixes #679
Code looks good to me; have you tested it? If not, I'll give it a try when I get the chance.
Once the --global glitch is fixed it'll be easier to test. Then, you could even add a unit test.
I'm thinking I should still fix the one remaining race condition; expect another push here soon (but I have meetings for several hours no, so "soon" is later than I'd like).
... Oh, it's merged now. https://github.com/Charcoal-SE/SmokeDetector/pull/694
What 'race conditions' remain? If you refer to the 'deploy' branch, that must be returned to before we can do anything else - no way around that - that is, during the 'lock' we need to be able to return to the 'deploy' branch as normal; all merges should be based on that after all.
So, please explain 'race conditions' in context, I don't see your 'race condition' here unfortunately.
Returning to the deploy branch currently happens outside the locked critical section, which means it could (attempt to) overwrite the work directory of another instance which is currently inside the critical section.
Codecov Report
Merging #695 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #695 +/- ##
=======================================
Coverage 61.65% 61.65%
=======================================
Files 5 5
Lines 1090 1090
=======================================
Hits 672 672
Misses 418 418
Impacted Files
Coverage ฮ
findspam.py
81.71% <0%> (รธ)
:arrow_up:
globalvars.py
92.98% <0%> (รธ)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
ฮ = absolute <relative> (impact), รธ = not affected, ? = missing data
Powered by Codecov. Last update 5d27c8e...1dc9170. Read the comment docs.
Also for the record, I described the race condition in more detail in the related issue https://github.com/Charcoal-SE/SmokeDetector/issues/679#issuecomment-299099042
The design of gitmanager.py is such that it is pretty much impossible to test it without complex, elaborate, and error-prone arrangements to have the master and deploy branches refer to something else than the real thing. I would hope this could be merged to Zephyr or something for testing if you are unwilling to merge it without some sort of system testing.
I'm happy for this to be merged, and if Something Bad Happensโข we can always revert it and try again.
I'd be happy to merge this and revert it, but firstly, let's merge in my Windows git-compatibility layer, though it should have no impact on this, i'd like to get that in and do testing on it and improvements on it, so that that isn't impacted by this directly. It shouldn't affect any actual Smokey instances, but testing is hard because it reverts to the master and deploy branches then dies on Windows (part of my 'Make it Windows Compatible' evil heh)
A quick test run of this functionality is demonstrated here:
http://chat.stackexchange.com/transcript/message/37345341#37345341
I challenge you to rapidly blacklist two entries with the current code base to see if it works.
Something was funky with the previous code. I have force-pushed a single commit which just adds the lock and wraps the critical region in a try: acquire ... finally: release
|
2025-04-01T04:10:16.189584
| 2022-12-17T18:43:15
|
1501523452
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13764",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/7406"
}
|
gharchive/pull-request
|
Ethan: Watch worldnewsrecords.com
Ethan requests the watch of the watch_keyword worldnewsrecords\.com. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
worldnewsrecords\.com has been seen in 1 true positive, 0 false positives, and 0 NAAs.
Approved by cocomac in Charcoal HQ
|
2025-04-01T04:10:16.192991
| 2023-12-19T15:32:24
|
2048900765
|
{
"authors": [
"SmokeDetector",
"metasmoke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13765",
"repo": "Charcoal-SE/SmokeDetector",
"url": "https://github.com/Charcoal-SE/SmokeDetector/pull/9521"
}
|
gharchive/pull-request
|
Jesse: Watch techsolidity.com
Jesse requests the watch of the watch_keyword techsolidity\.com. See the MS search here and the Stack Exchange search in text, in URLs, and in code.
techsolidity\.com has been seen in 1 true positive, 0 false positives, and 0 NAAs.
Approved by Jeff Schaller in Charcoal HQ
|
2025-04-01T04:10:16.208451
| 2017-10-30T01:53:44
|
269460053
|
{
"authors": [
"TheMangoTrain",
"danersan",
"zixia"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13766",
"repo": "Chatie/wechaty",
"url": "https://github.com/Chatie/wechaty/issues/936"
}
|
gharchive/issue
|
[CI] Homebrew must be run under Ruby 2.3! You're running 2.0.0. (RuntimeError)
https://github.com/travis-ci/travis-ci/issues/8552
TravisCI MacOS Error:
/usr/local/Homebrew/Library/Homebrew/brew.rb:12:in `<main>': Homebrew must be run under Ruby 2.3! You're running 2.0.0. (RuntimeError)
https://travis-ci.org/Chatie/wechaty/jobs/294655691#L3326
Run brew uninstall --force brew-cask will prevent this error.
Forgot the reason for why...
Thanks, worked for me. Had same error as OP.
crap, that's worked for me, thanks buddy
|
2025-04-01T04:10:16.214943
| 2021-03-27T04:27:36
|
842413621
|
{
"authors": [
"MrAuro",
"jammehcow",
"pajlada"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13767",
"repo": "Chatterino/chatterino2",
"url": "https://github.com/Chatterino/chatterino2/issues/2568"
}
|
gharchive/issue
|
When opening a modified emote's TwitchEmotes link the page can't be found
Describe the bug
The TwitchEmotes.com link says "Sorry, the page you are looking for could not be found." when opening a modified emote.
To reproduce
Right click and open the TwitchEmotes link of a modified emote
Screenshots
Modified emote, pokeWICKED with the horizontal flip modification:
Opening the TwitchEmote link gives this error:
Chatterino version
Chatterino Nightly 2.2.3-beta2 (commit 1f5b62e6e)
Operating system
Windows 10
Additional information
Removing the modification addon like _HF from the URL works fine
Can reproduce in Chatterino 2.2.3-beta2 (commit 1f5b62e6e). Note that this is an issue with TwitchEmotes themselves as it doesn't seem like they store modifications (understandably). I wonder if it's worth just sending the user to the base emote's page for now? Thoughts @pajlada @fourtf ?
Message posting the pajaW sub emote
@badge-info=subscriber/63;badges=broadcaster/1,subscriber/3060,partner/1;color=#CC44FF;display-name=pajlada;emote-only=1;emotes=80481:0-4;flags=;id=3fc46182-c105-4666-8af5-60a1aa7adcef;mod=0;room-id=11148817;subscriber=1;tmi-sent-ts=1616937942185;turbo=0;user-id=11148817;user-type=<EMAIL_ADDRESS>PRIVMSG #pajlada :pajaW
Message posting the pajaW sub emote with the sunglasses emote modifier
@badge-info=subscriber/63;badges=broadcaster/1,subscriber/3060,partner/1;color=#CC44FF;display-name=pajlada;emote-only=1;emotes=80481_SG:0-7;flags=;id=feb420da-e42e-4fa1-bf9e-cced8d0d3b6f;mod=0;room-id=11148817;subscriber=1;tmi-sent-ts=1616937943809;turbo=0;user-id=11148817;user-type=<EMAIL_ADDRESS>PRIVMSG #pajlada :pajaW_SG
Message posting the pajaPants bit emote
@badge-info=subscriber/63;badges=broadcaster/1,subscriber/3060,partner/1;color=#CC44FF;display-name=pajlada;emote-only=1;emotes=emotesv2_56124b6ee48c47778de1a922e93d7848:0-8;flags=;id=f7d415ac-f088-4577-9bc6-a91b280a2ffc;mod=0;room-id=11148817;subscriber=1;tmi-sent-ts=1616937999027;turbo=0;user-id=11148817;user-type=<EMAIL_ADDRESS>PRIVMSG #pajlada :pajaPants
Note the various emote keys in use right now:
80481
80481_SG
emotesv2_56124b6ee48c47778de1a922e93d7848
Any sort of "trying to be smart" parsing from our end will most likely not hold the test of time, so the way I see it there's not much we can do
|
2025-04-01T04:10:16.218388
| 2024-04-21T17:12:51
|
2255179416
|
{
"authors": [
"core-code",
"pajlada"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13768",
"repo": "Chatterino/chatterino2",
"url": "https://github.com/Chatterino/chatterino2/issues/5344"
}
|
gharchive/issue
|
2.5.0 is a pre-release or stable?
Checklist
[X] I'm reporting a problem with Chatterino
[X] I've verified that I'm running the most recent nightly build or stable release
[X] I've looked for my problem on the wiki
[X] I've searched the issues and pull requests for similar looking reports
Describe your issue
i am confused about 2.5.0 - on github it is marked as a prerelease, on chatterino.com it is the default download
Screenshots
none
OS and Chatterino Version
any
Fixed, thanks. 2.5.0 is the latest stable release
|
2025-04-01T04:10:16.242981
| 2024-03-16T00:01:49
|
2189630089
|
{
"authors": [
"Tellinq",
"kezukdev"
],
"license": "CC0-1.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13769",
"repo": "CheatBreakerNet/Client-API",
"url": "https://github.com/CheatBreakerNet/Client-API/pull/82"
}
|
gharchive/pull-request
|
Added akyto
General:
[x] The pull request title is descriptive. (ex. Added GoldenPvP Network, not Updated metadata.json)
[x] The pull request does not contain any unrelated commits. (ex. commits from previous pull requests)
Mapping Additions or Updates:
[x] Folder with appropriate name has been created / updated (must not include any non-alpha characters or whitespace).
[x] No changes were made to the file's formatting (4 spaces per tab).
[x] There are no syntax errors.
[x] There is no pre-existing mapping matching my id (check if there is an existing folder).
[x] My field values match your requirements.
You can view our patterns here: metadata.schema.json, or take a look below and complete the field checklist:
[x] id: a lowercase string, which should match the folder name (ex. goldenpvpnetwork)
[x] name: a string (ex. GoldenPvP Network)
[x] description: hook between 16 and 40 characters (ex. We're the home of classic PvP gamemodes)
[x] addresses: an array with lowercase strings (ex. ["goldenpvp.net", "routing.center"])
You do not need to specify subdomains, CheatBreaker services automatically detect them.
[x] primaryAddress: the primary address that people connect to with (please include the subdomain if required) (ex. mc.goldenpvp.net)
[x] minecraftVersions: an array with Minecraft versions as strings (ex. ["1.7.*", "1.8.9"] - Must be versions or subversions supported by CheatBreaker)
[x] primaryMinecraftVersion: a Minecraft version as a string (ex. 1.7.10 - Must be a subversion supported by CheatBreaker)
[x] primaryColor: a hexademical color code that primarily distinguishes the server (ex. #00FFFF)
[x] secondaryColor: a hexademical color code that accompanies the primaryColor of the server (ex. #FF0000)
[x] primaryRegion: the primary region where your server operates in (ex. EU)
[x] regions: a list of regions where you have servers located that service your players (ex. ["EU", "NA", "AS"])
[x] gameTypes: a list of games that describe the content on your server, must be a max of 3 listed (ez. ["PVP", "UHC", "HCF"])
[ ] (optional) website: url of server website, must include URL schema (http:// or https://) (ex. https://www.goldenpvp.net)
[ ] (optional) store: url of server store, must include URL schema (http:// or https://) (ex. https://store.goldenpvp.net)
Socials
[ ] (optional) twitter: username of twitter account without the @ (ex. GoldenPvPNet)
[x] (optional) discord: invite link to discord (ex. https://discord.com/invite/3EAzBxf)
[ ] (optional) youtube: slug or username of youtube channel (ex. GoldenPvP)
[ ] (optional) instagram: username of instagram account (ex. GoldenPvPNet)
[ ] (optional) twitch: username of twitch account (ex. GoldenPvPNet)
[ ] (optional) telegram: slug of telegram group (ex. GoldenPvP-Network)
[ ] (optional) reddit: slug of subdreddit wtih 'r/' (ex. GoldenPvP)
[ ] (optional) tiktok: username of tiktok account (ex. GoldenPvP)
[ ] (optional) facebook: slug of facebook page (ex. GoldenPvP)
Media:
Icon
[x] My icon is a png file.
[x] I have uploaded my icon to my server folder (ex. goldenpvpnetwork) and named it icon.png.
[x] My icon has a transparent background and is square (1:1 aspect ratio).
[x] My icon is 64 pixels in width and height.
[x] My icon is my own/the server's property and complies with relevant copyright/privacy laws.
Logo
[x] My logo is a png file.
[x] I have uploaded my logo to my server folder (ex. goldenpvpnetwork) and named it logo.png.
[x] My logo has a transparent background and is square (1:1 aspect ratio).
[x] My logo is 108 pixels in width and height.
[x] My logo is my own/the server's property and complies with relevant copyright/privacy laws.
Discord Logo
[x] My logo is a png file.
[x] I have uploaded my logo to my server folder (ex. goldenpvpnetwork) and named it discord.png.
[x] My logo has a transparent background and is square (1:1 aspect ratio).
[x] My logo is between 512 and 1024 pixels in width and height.
[x] My logo is my own/the server's property and complies with relevant copyright/privacy laws.
Primary Background (Only required for partnered servers)
[x] My primary background is a png file.
[x] I have uploaded my primary background to my server folder (ex. goldenpvpnetwork) and named it primarybackground.png.
[x] My primary background is 774 pixels in width and 363 pixels in height.
[x] My primary background contains relevant content pertaining to the server, is my own/the server's property, and complies with relevant copyright/privacy laws.
[x] My primary background doesn't contain any markings, text, or logos (unless part of a build).
Secondary Background (Only required for partnered servers)
[x] My secondary background is a png file.
[x] I have uploaded my secondary background to my server folder (ex. goldenpvpnetwork) and named it secondarybackground.png.
[x] My secondary background is 447 pixels in width and 172 pixels in height.
[x] My secondary background contains relevant content pertaining to the server, is my own/the server's property, and complies with relevant copyright/privacy laws.
[x] My secondary background doesn't contain any markings, text, or logos (unless part of a build).
Hold on...
I've seen this specific description before. Did you by any chance rip this from another metadata file as a template but not change it?
I cannot accept this PR due to the logos violating copyright usage. Even so, they miss our transparent requirement.
https://tineye.com/search/79b55ab38b04927adac736ef338f47a0b2c81adc?sort=score&order=desc&page=1
Please understand that while it might be a slight nuance for you, copyright is a serious matter for us and we do not accept mappings that contain copyright infringement.
I unchecked the tasks that are required for partnered servers only. Since you did not supply any, do not check these. If you wish to fill out the requirements, make sure you include images that would meet the visual requirements.
|
2025-04-01T04:10:16.251517
| 2022-05-01T01:49:08
|
1221977250
|
{
"authors": [
"kaplanlior",
"tspearconquest"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13770",
"repo": "Checkmarx/kics",
"url": "https://github.com/Checkmarx/kics/pull/5282"
}
|
gharchive/pull-request
|
Typo fix
Signed-off-by: Thomas Spear<EMAIL_ADDRESS>Proposed Changes
Fixes a typo in dockerfile missing dnf flag query
I submit this contribution under the Apache-2.0 license.
Thanks for catching this (:
The documentation actually says there isn't any hyphen there.
Can you change that as well ?
See https://dnf.readthedocs.io/en/latest/command_ref.html#options
|
2025-04-01T04:10:16.254013
| 2024-07-02T10:01:42
|
2385815481
|
{
"authors": [
"AlvoBen",
"pedrompflopes"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13771",
"repo": "CheckmarxDev/ast-cli-java-wrapper",
"url": "https://github.com/CheckmarxDev/ast-cli-java-wrapper/pull/334"
}
|
gharchive/pull-request
|
Update checkmarx-ast-cli binaries with 2.1.6
Updates checkmarx-ast-cli to 2.1.6
Auto-generated by [create-pull-request][2]
Checkmarx One โ Scan Summary & Details โ 59066765-884a-468f-b6e8-46be79b4355a
No New Or Fixed Issues Found
|
2025-04-01T04:10:16.301079
| 2022-04-05T12:35:11
|
1193089236
|
{
"authors": [
"NaridaL",
"bd82",
"mattbishop"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13772",
"repo": "Chevrotain/chevrotain",
"url": "https://github.com/Chevrotain/chevrotain/pull/1790"
}
|
gharchive/pull-request
|
copy StringLiteral from benchmark parser
See https://github.com/Chevrotain/chevrotain/blob/master/packages/chevrotain/benchmark_web/parsers/json/json_parser.js
":?" is an optional :, not a non-capturing group. Additionally, nested + and * lead to terrible performance when not matching.
Fixes #1788
Json.org has a bnf describing valid chars. Just use that?
@mattbishop is correct, we should re-evaluate the regexp from scratch, It is possible I copy pasted
it in the past and that it is not fully aligned with the JSON spec.
It may not be a direct 1 to 1 translation as the BNF may not translate to JavaScript regexp directly and JS string quirks.
|
2025-04-01T04:10:16.391160
| 2020-04-21T19:15:55
|
604221341
|
{
"authors": [
"PascalSenn",
"hotgen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13773",
"repo": "ChilliCream/hotchocolate",
"url": "https://github.com/ChilliCream/hotchocolate/issues/1745"
}
|
gharchive/issue
|
IQueryable in resolver: random errors
Vesrsion: 10.4, EF Core 3.1
If I return IQueryable from a resolver I got random errors. If I switch to List everything fine.
I have a many to many relationship and class like this:
public class Card
{
public int Id { get; set; }
public DateTime CreatedAt { get; set; }
public string Name { get; set; }
public string Annot { get; set; }
public IQueryable<File> GetFiles([Service]Database db)
{
return db.FileCards
.Include(fc => fc.File)
.Where(fc => fc.CardId == Id)
.Select(fc => fc.File);
}
}
Query: {cards{name, files{name}}}
Output (changing each time):
{
"errors": [
{
"message": "Unexpected Execution Error",
"locations": [
{
"line": 4,
"column": 5
}
],
"path": [
"cards",
0,
"files"
]
},
{
"message": "Unexpected Execution Error",
"locations": [
{
"line": 4,
"column": 5
}
],
"path": [
"cards",
1,
"files"
]
}
],
"data": {
"cards": [
{
"name": "Card1",
"files": null
},
{
"name": "Card2",
"files": null
},
{
"name": "Card3",
"files": []
}
]
}
}
With List (correct):
{
"data": {
"cards": [
{
"name": "Card1",
"files": [
{
"name": "File2"
},
{
"name": "File1"
}
]
},
{
"name": "Card2",
"files": []
},
{
"name": "Card3",
"files": []
}
]
}
}
Is it a BUG or I am doing something wrong?
Hi @hotgen
Thank you for reporting this error!
Did you configure ForceSerialExecution?
For that replace the ConfigureServices method with the following code.
public void ConfigureServices(IServiceCollection services)
{
services.AddDbContext<SchoolContext>();
services.AddGraphQL(
SchemaBuilder.New()
.AddQueryType<Query>()
.Create(),
new QueryExecutionOptions { ForceSerialExecution = true });
}
The difference between IQueryable and List is that one is resolved in memory and the other one is projected directly to your database.
As a sidenote:
You probably do not need this :)
public IQueryable<File> GetFiles([Service]Database db)
{
return db.FileCards
.Include(fc => fc.File)
.Where(fc => fc.CardId == Id)
.Select(fc => fc.File);
}
Because when you add [UseSelection] to your query resolver like this:
public class Query
{
[UseSelection]
[UseFiltering]
[UseSorting]
public IQueryable<Card> GetCards([Service]Database context) =>
context.Cards;
}
and have a relation tho your Cards on the mode like this:
public class Card
{
public int Id { get; set; }
public DateTime CreatedAt { get; set; }
public string Name { get; set; }
public string Annot { get; set; }
public List<File> Files {get;set; }
}
the query is projected to the database and includes all the fields.
Given a query like
query {
cards(where: { OR: [{ name: "Bar" }, { annot: "Baz" }] }) {
id
name
files {
name
}
}
}
Creates an equivalent sql to the following statement:
SELECT "c"."Id",
"c"."Name",
"t"."Name",
FROM "Cards" AS "c"
LEFT JOIN (
SELECT "f"."Name",
FROM "Files" AS "f"
) AS "t" ON "f"."Id" = "t"."CardId"
WHERE ("s"."Name" = 'Bar') OR ("s"."Annot" = 'Baz')
This would also work with nested filtering and sorting
Check out this blog post for more information:
https://dev.to/michaelstaib/get-started-with-hot-chocolate-and-entity-framework-e9i
@PascalSenn Very helpful. Thank you!
Your are most welcome :)
Can we close this one?
|
2025-04-01T04:10:16.397771
| 2021-05-26T20:27:38
|
902884962
|
{
"authors": [
"Zeitwaechter"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13774",
"repo": "ChimeraOS/chimeraos",
"url": "https://github.com/ChimeraOS/chimeraos/issues/231"
}
|
gharchive/issue
|
Unable to login via Terminal after GamerOS upgrade from v23 > v24
:wave:
So, to recap from my (sadly) hijacking in #212.
General approach to give an understanding of my stituation:
I can't really login via termial anymore as I'm directly forwarded back to the terminal's login.
I can sometimes see that the login worked, for a split second.. Pressing CTRL+C did not work, already checked.
Also no luck with uploading my ED25519 key via Steam Buddy, so I can login via SSH.
ssh -v from my 2nd system has shown me that the key is introduced to the GamerOS system, but not accepted.
Not so sure what is happening here, as ED25519 is also in the list of acceptable keys.
Now the more procedual approach:
Installed GamerOS 23 weeks before
Could login with gamer / gamer
Did not change the login credentials
Unlocked frzr via sudo frzr-unlock (here we go boiiis)
Did not switch from stable to testing
Probably updated packages via pacman, sometime prior to the v24 upgrade
As the service to auto-update is always failing for me, on OS boot, I did start it manually
Rebooted the systemd after the started service had finished (new image downloaded) + systemctl status frzr-autoupdate-something did say so (sorry for not getting the name on point)
Could not login anymore
Uploaded SSH key of a 2nd system and uploaded it to gameros.local / the devices IP / Steam Buddy
Could not login via SSH, even though the key was introduced withing ssh -v log.
As matters are I cannot verify anything on the maschine, using regular login.
I could surely access the SSD via my main Arch Linux, but I had once memory corruption due to bad mounting / block resizing, when doing so.. so I'ld rather not be tempted to do so.. but I guess there is no way around it to get to the bottom of this, is there? :S
Let me know if and how I can be of further help to you :+1:
Kind regards
Edit (2021-06-15):
Can now also confirm, SSH Key is actually correctly imported.
JFYI: This did not happen to me on v24 > v25, a few weeks ago.
Yeah, for now :)
|
2025-04-01T04:10:16.399988
| 2022-11-28T15:18:46
|
1466610487
|
{
"authors": [
"houku",
"pastaq"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13775",
"repo": "ChimeraOS/chimeraos",
"url": "https://github.com/ChimeraOS/chimeraos/issues/438"
}
|
gharchive/issue
|
a good amount of bugs here
so i recently installed chimera on an asus vivobook x515ma and aya neo 2021, and the installer was good and all,
however, when i went to access the gamepadui, the screen was flikering on the laptop, and when i changed to desktop, the screen was flikering still, but changing the display resolution seemed to fix it, and theres no flikering in my aya neo in gamepad
another bug is external storage, i connected an external hard drive and its found in gnome disks, but not in gnome files, so i cant transfer my roms and such if i want to use retrodeck, (i know i can use the chimera web app, but its an example), installing a different file manager on laptop didnt seem to find the external storage, my aya neo on gamepadui doesnt seem to find external storage either
flatpak seems to work fine, and i do like how minimal the desktop is,
game mode on desktop seems to get on a blank screen on the laptop, but i didnt stay long since i wanted to get desktop stuff working first, on the aya neo 2021, gamepadui gets eventually to the steam deck ui,
Thank you for reporting these issues. Can you open a separate issue for each distinct bug? It will be much easier to troubleshoot and track progress for us in resolving them.
|
2025-04-01T04:10:16.460534
| 2024-05-16T02:08:12
|
2299175146
|
{
"authors": [
"ChrisDill",
"i3elj"
],
"license": "Zlib",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13776",
"repo": "ChrisDill/Raylib-cs",
"url": "https://github.com/ChrisDill/Raylib-cs/issues/241"
}
|
gharchive/issue
|
Incorrect Namespacing in Camera2DDemo Example and probably others
Issue description
After creating a blank project using dotnet new console and adding Raylib-cs with dotnet add package Raylib-cs, the code sample provided here doesn't work out of the box which I imagine that it should (I could be wrong and please point out if I am. I'm new to C# and I don't really know how things work just yet)
Environment and other useful information
OS: Linux (Arch Linux)
Mesa drives version: 1:24.0.7-3
Raylib version: 5.0-1
Raylib-cs version: 6.0.0
Reproducing errors
mkdir test/
cd test/
dotnet new console
dotnet add package Raylib-cs
copy code from Examples/Core/Camera2dDemo.cs into the generated Program.cs. Run:
dotnet run
Expected output
INFO: Initializing raylib 5.0
INFO: Platform backend: DESKTOP (GLFW)
INFO: Supported raylib modules:
INFO: > rcore:..... loaded (mandatory)
INFO: > rlgl:...... loaded (mandatory)
INFO: > rshapes:... loaded (optional)
INFO: > rtextures:. loaded (optional)
INFO: > rtext:..... loaded (optional)
INFO: > rmodels:... loaded (optional)
INFO: > raudio:.... loaded (optional)
INFO: DISPLAY: Device initialized successfully
INFO: > Display size: 1920 x 1080
INFO: > Screen size: 800 x 450
INFO: > Render size: 800 x 450
INFO: > Viewport offsets: 0, 0
INFO: GLAD: OpenGL extensions loaded successfully
INFO: GL: Supported extensions count: 230
INFO: GL: OpenGL device information:
INFO: > Vendor: Intel
INFO: > Renderer: Mesa Intel(R) UHD Graphics (TGL GT2)
INFO: > Version: 4.6 (Core Profile) Mesa 24.0.7-arch1.3
INFO: > GLSL: 4.60
INFO: GL: VAO extension detected, VAO functions loaded successfully
INFO: GL: NPOT textures extension detected, full NPOT textures supported
INFO: GL: DXT compressed textures supported
INFO: GL: ETC2/EAC compressed textures supported
INFO: PLATFORM: DESKTOP (GLFW): Initialized successfully
INFO: TEXTURE: [ID 1] Texture loaded successfully (1x1 | R8G8B8A8 | 1 mipmaps)
INFO: TEXTURE: [ID 1] Default texture loaded successfully
INFO: SHADER: [ID 1] Vertex shader compiled successfully
INFO: SHADER: [ID 2] Fragment shader compiled successfully
INFO: SHADER: [ID 3] Program shader loaded successfully
INFO: SHADER: [ID 3] Default shader loaded successfully
INFO: RLGL: Render batch vertex buffers loaded successfully in RAM (CPU)
INFO: RLGL: Render batch vertex buffers loaded successfully in VRAM (GPU)
INFO: RLGL: Default OpenGL state initialized successfully
INFO: TEXTURE: [ID 2] Texture loaded successfully (128x128 | GRAY_ALPHA | 1 mipmaps)
INFO: FONT: Default font loaded successfully (224 glyphs)
INFO: TIMER: Target time per frame: 16.667 milliseconds
INFO: TEXTURE: [ID 2] Unloaded texture data from VRAM (GPU)
INFO: SHADER: [ID 3] Default shader unloaded successfully
INFO: TEXTURE: [ID 1] Default texture unloaded successfully
INFO: Window closed successfully
Actual output and errors
/path/to/project/Program.cs(30,9): error CS0246: The type or namespace name 'Rectangle' could not be found (are you missing a using directive or an assembly reference?) [/path/to/project/temp.csproj]
/path/to/project/Program.cs(31,9): error CS0246: The type or namespace name 'Rectangle' could not be found (are you missing a using directive or an assembly reference?) [/path/to/project/temp.csproj]
/path/to/project/Program.cs(31,37): error CS0246: The type or namespace name 'Rectangle' could not be found (are you missing a using directive or an assembly reference?) [/path/to/project/temp.csproj]
/path/to/project/Program.cs(32,9): error CS0246: The type or namespace name 'Color' could not be found (are you missing a using directive or an assembly reference?) [/path/to/project/temp.csproj]
/path/to/project/Program.cs(32,35): error CS0246: The type or namespace name 'Color' could not be found (are you missing a using directive or an assembly reference?) [/path/to/project/temp.csproj]
/path/to/project/Program.cs(45,34): error CS0246: The type or namespace name 'Color' could not be found (are you missing a using directive or an assembly reference?) [/path/to/project/temp.csproj]
/path/to/project/Program.cs(53,9): error CS0246: The type or namespace name 'Camera2D' could not be found (are you missing a using directive or an assembly reference?) [/path/to/project/temp.csproj]
/path/to/project/Program.cs(69,27): error CS0103: The name 'KeyboardKey' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(73,32): error CS0103: The name 'KeyboardKey' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(82,27): error CS0103: The name 'KeyboardKey' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(86,32): error CS0103: The name 'KeyboardKey' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(114,30): error CS0103: The name 'KeyboardKey' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(124,29): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(128,52): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(135,38): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(137,83): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(143,17): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(148,50): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(150,54): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(151,60): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(152,79): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(153,74): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(155,56): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(156,50): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(158,62): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(159,65): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(160,66): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(161,55): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
/path/to/project/Program.cs(162,69): error CS0103: The name 'Color' does not exist in the current context [/path/to/project/temp.csproj]
The build failed. Fix the build errors and run again.
Temporary fix
I managed to make it work by changing using static Raylib_cs.Raylib to using Raylib_cs and adding Raylib. at the beginning of some types. Not all of them needed tho, Rectangle, Camera2D and Color are examples that only worked without the namespace.
After reading some old closed issues I found that by adding using Raylib_cs on top of using static Raylib_cs.Raylib fixed too.
@i3elj Yeah it needs using Raylib_cs which the examples currently sets in Examples.csproj as a global using so each example doesn't need to specify it to have access to the namespace.
|
2025-04-01T04:10:16.481264
| 2021-09-28T09:20:10
|
1009499323
|
{
"authors": [
"NekoWings",
"ZACK741",
"zeykafx"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13777",
"repo": "ChrisKnott/Eel",
"url": "https://github.com/ChrisKnott/Eel/issues/525"
}
|
gharchive/issue
|
React with Eel
Describe the problem
I'm trying to merge React with Eel in Eel's GUI
React requires an eel object while building, Wonder is there any way to allow React to skip this.
I've tried example 07, but it seems to run on the browser only.
Is there any way to achieve this?
you can use this https://github.com/hrdtbs/template-create-react-app-eel
I managed to get electron+react+eel working thanks to that template.
Thanks to the template. But I've got some problem.
How did you import eel.js in react part.
I've got Failed to load resource: net::ERR_CONNECTION_REFUSED when using <script type="text/javascript" src="http://localhost:8080/eel.js"></script>
And i've got Uncaught SyntaxError: Unexpected token '<' when using <script type="text/javascript" src="/eel.js"></script>
Or is there a correct setting for the eel.start options in python part?
|
2025-04-01T04:10:16.488453
| 2024-11-22T15:29:08
|
2683567546
|
{
"authors": [
"OvermindDL1",
"adamperkowski"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13779",
"repo": "ChrisTitusTech/linutil",
"url": "https://github.com/ChrisTitusTech/linutil/issues/963"
}
|
gharchive/issue
|
Cargo installation instructions incorrect
Describe the bug
The cargo installation instructions state to run cargo install linutil but that is incorrect, it should be cargo install linutil_tui. (See additional notes below as well.)
To Reproduce
Steps to reproduce the behavior:
Look at the crates.io page. Try copy pasting the cargo install line and run it.
Expected behavior
Copy pasting the command to install should work
Checklist
[X] I checked for duplicate issues.
Additional notes:
The version currently on crates.io does not compile anyway because someone broke semver. You actually need to run cargo install linutil_tui --locked or it won't compile.
hmm... i wonder who wrote that (it was me)
|
2025-04-01T04:10:16.492245
| 2024-11-19T02:11:59
|
2670498338
|
{
"authors": [
"jeevithakannan2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13780",
"repo": "ChrisTitusTech/linutil",
"url": "https://github.com/ChrisTitusTech/linutil/pull/956"
}
|
gharchive/pull-request
|
refact(installer, updater): formatting & small fixes
Type of Change
[x] Bug fix
[x] Refactoring
Description
Installer -> Remove unreachable pacman case. dnf requires rustup-init before using rustup-init. Source .cargo/env after the installation of rustup.
Updater -> Formatting, Install rustup from dnf, change the wild case to use rustup.rs general installer from rust. Source the .cargo/env after the installation of rustup.
Testing
Tested on fedora 41 without issues.
Issues / other PRs related
Resolves #955
Checklist
[x] My code adheres to the coding and style guidelines of the project.
[x] I have performed a self-review of my own code.
[x] I have commented my code, particularly in hard-to-understand areas.
[x] My changes generate no errors/warnings/merge conflicts.
No until you do rustup-init rustup command will not be available in fedora and alpine
|
2025-04-01T04:10:16.500773
| 2024-02-06T16:32:15
|
2121217477
|
{
"authors": [
"AlexMayne5",
"possumdoe",
"shrimp332",
"ubrider"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13781",
"repo": "ChrisTitusTech/winutil",
"url": "https://github.com/ChrisTitusTech/winutil/issues/1580"
}
|
gharchive/issue
|
The Raspberry Pi Imager is not installing
Describe the bug
A clear and concise description of what the bug is.
The Raspberry Pi Imager is not installing with error and got the same results after reinstall of winget
To Reproduce
Steps to reproduce the behavior:
reinstall winget using button
on verstion 24.02.03 of the script click the check box for Raspberry Pi Imager
click install Selected
see the error
Expected behavior
uses WinGet to install Raspberry Pi Imager
Screenshots
If applicable, add screenshots to help explain your problem.
Additional context
Add any other context about the problem here.
N/A
Hi i got the same on win 11 23h2 for whatever package...
In the winget-pkgs repository the package RaspberryPiFoundation.RaspberryPiImager's manifest doesn't specify the installation scope.
In the script the line that installs the packages is the following:
winget install -e --accept-source-agreements --accept-package-agreements --scope=machine --silent RaspberryPiFoundation.RaspberryPiImager
So maybe the restriction to --scope=machine cause RaspberryPiFoundation.RaspberryPiImager installation to fail.
The same error occurs in the issue #1494 with TorProject.TorBrowser, SomePythonThings.WingetUIStore, Nvidia.GeForceNow which do NOT specify the installer for the machine scope.
Other apps, like Google.Chrome or Mozilla.Firefox, specify the installer for the machine scope.
You can see which packages contains this by using the Github's search functionality:
This is the search string I used:
repo:microsoft/winget-pkgs path:/^manifests\/.*installer.yaml/ "Scope: machine" Package.Name
Imho this is the reason why the package will fail to install through the script.
nepnep.neofetch-win also has this problem
|
2025-04-01T04:10:16.516818
| 2021-09-30T09:06:56
|
1011918939
|
{
"authors": [
"ChristianMarzahl",
"maubreville"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13782",
"repo": "ChristianMarzahl/Exact",
"url": "https://github.com/ChristianMarzahl/Exact/issues/66"
}
|
gharchive/issue
|
Annotations being only partially saved if not hit
When adding rectangular annotations (only checked with those) within an image, these are only reliably stored into the database when "enter" is hit. While this is the intended workflow, the annotations are shown on the UI regardless if enter was hit or not.
In the following image, the left annotations were all clicked only a single time (and are shown all), without hitting enter at any time:
This is the result in the database:
Since the annotations are successfully shown, it is not unlikely that users deviate from the instructions and don't hit enter.
I would encourage to show annotations only after successful storage in the database (i.e. on server response), to have DB and UI always in sync, if this is not already the case.
The slide (for reference) is: https://exact.cs.fau.de/annotations/19219/?frame=1&xmin=5272&ymin=3975&xmax=7182&ymax=5121
Dear User,
I would encourage to show annotations only after successful storage in the database (i.e. on server response), to have DB and UI always in sync, if this is not already the case.
There is a notification at the bottom if the annotation was created successfully.
Fixed in version: https://github.com/ChristianMarzahl/Exact/pull/67/
|
2025-04-01T04:10:16.527049
| 2024-08-10T08:45:59
|
2458991948
|
{
"authors": [
"VijaySamant4368",
"ayush-t02"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13783",
"repo": "ChromeGaming/Dot-Box",
"url": "https://github.com/ChromeGaming/Dot-Box/pull/758"
}
|
gharchive/pull-request
|
Added testimonials
PR Description ๐
Added testimonials navigation link to the following pages:
how-to-play page
about page
FAQ page
contributors page
testimonials page
Fixes #754
Mark the task you have completed โ
[x] I follow CONTRIBUTING GUIDELINE & CODE OF CONDUCT of this project.
[x] I have performed a self-review of my own code or work.
[x] I have commented my code, particularly in hard-to-understand areas.
[x] My changes generates no new warnings.
[x] I have followed proper naming convention showed in CONTRIBUTING GUIDELINE
[x] I have added screenshot for review.
## Add your screenshots(Optional) ๐ธ
Thank you soo much for contributing to our repository ๐
@ayush-t02 please check
Resolve conflicts
@ayush-t02 resolved.
|
2025-04-01T04:10:16.538630
| 2021-12-15T07:45:03
|
1080687881
|
{
"authors": [
"ChunghaSung",
"LuckyYZC"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13784",
"repo": "ChunghaSung/intAbs",
"url": "https://github.com/ChunghaSung/intAbs/issues/1"
}
|
gharchive/issue
|
Bugs I encountered when running IntAbs
Dear Author,
I encountered some bugs when I running IntAbs. If you have some suggestions for fixing these errors, I would be very grateful.
When I run the instruction "make", it reports the following errors.
[ 16%] Building CXX object worklist-ai/CMakeFiles/worklistAI.dir/worklistai.cpp.o
/home/guisen/source/intAbs/src/worklist-ai/worklistai.cpp: In member function โvoid WorklistAI::addDomLoadStoreFacts(std::set<llvm::Function*>, z3::context&, _Z3_fixedpoint*&)โ:
/home/guisen/source/intAbs/src/worklist-ai/worklistai.cpp:2038:45: error: no matching function for call to โllvm::DominatorTree::DominatorTree(llvm::Function&)โ
DominatorTree &dt = DominatorTree(*f);
^
/home/guisen/source/intAbs/src/worklist-ai/worklistai.cpp:2038:45: note: candidates are:
In file included from /usr/lib/llvm-3.6/include/llvm/Analysis/PostDominators.h:17:0,
from /home/guisen/source/intAbs/src/worklist-ai/worklistai.cpp:20:
/usr/lib/llvm-3.6/include/llvm/IR/Dominators.h:77:3: note: llvm::DominatorTree::DominatorTree(llvm::DominatorTree&&)
DominatorTree(DominatorTree &&Arg)
^
/usr/lib/llvm-3.6/include/llvm/IR/Dominators.h:77:3: note: no known conversion for argument 1 from โllvm::Functionโ to โllvm::DominatorTree&&โ
/usr/lib/llvm-3.6/include/llvm/IR/Dominators.h:75:3: note: llvm::DominatorTree::DominatorTree()
DominatorTree() : DominatorTreeBase<BasicBlock>(false) {}
^
/usr/lib/llvm-3.6/include/llvm/IR/Dominators.h:75:3: note: candidate expects 0 arguments, 1 provided
make[2]: *** [worklist-ai/CMakeFiles/worklistAI.dir/worklistai.cpp.o] Error 1
make[1]: *** [worklist-ai/CMakeFiles/worklistAI.dir/all] Error 2
make: *** [all] Error 2
It looks like the function DominatorTree is being used incorrectly. So I replace "DominatorTree &dt = DominatorTree(*f);" by
"DominatorTree &dt = getAnalysis(*f).getDomTree();" in the worklistai.cpp. Then it works, the command โmakeโ can be run through. But when I use the libworklistAI.so built to run the cases in the directory test, it does not work, and reports the following errors.
guisen@ubuntu:~/source/intAbs/src/build$ make
Scanning dependencies of target worklistAI
[ 16%] Building CXX object worklist-ai/CMakeFiles/worklistAI.dir/worklistai.cpp.o
Linking CXX shared module libworklistAI.so
[100%] Built target worklistAI
guisen@ubuntu:~/source/intAbs/src/build$ cd ../test/
guisen@ubuntu:~/source/intAbs/src/test$ ./run.sh test1
Testing test1
/usr/lib/llvm-3.6/bin/clang -emit-llvm -S -c main.c
/usr/bin/time -f %e -o box_ncomb_nonpri.time /usr/lib/llvm-3.6/bin/opt -load /home/guisen/source/intAbs/src/build/worklist-ai/libworklistAI.so -worklist-ai -nocombs -box main.ll >main_out.bc 2>box_ncomb_nonpri.out
Tue Dec 14 23:35:05 PST 2021
139
[ERROR]: opt (box ncomb)
ERROR: get_num_errs(): error string not found in output
TEST PASSED: box_ncomb_nonpri
/usr/bin/time -f %e -o box_ncomb_pri.time /usr/lib/llvm-3.6/bin/opt -load /home/guisen/source/intAbs/src/build/worklist-ai/libworklistAI.so -worklist-ai -nocombs -priority -box main.ll >main_out.bc 2>box_ncomb_pri.out
Tue Dec 14 23:35:05 PST 2021
139
[ERROR]: opt (box ncomb)
ERROR: get_num_errs(): error string not found in output
TEST PASSED: box_ncomb_pri
Hello, I just found this post.
I think this is a version issue. APIs usages are different in different LLVM versions and I think I had similar problem when I worked on this (like several years ago)
Could you try to use LLVM 3.7 instead? Based on my commit history, I think I used 3.7, not 3.6.
Also, other way to create a domTree can be found here in LLVM 3.7: https://github.com/trailofbits/DomTreSat/blob/master/LLVMPassBuild/LLVMPassBuild/LLVMDomTreePass.cpp#L148
Please let me know if you still have this issue.
|
2025-04-01T04:10:16.565975
| 2024-04-13T17:17:28
|
2241663293
|
{
"authors": [
"Cidan",
"fiote"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13785",
"repo": "Cidan/BetterBags",
"url": "https://github.com/Cidan/BetterBags/pull/329"
}
|
gharchive/pull-request
|
Feature/move all from category
A really useful feature that Adibags had which seems to be missing from BetterBags.
Rightclicking section names on the backpack frame will (try to) move all its items to the bank.
Rightclicking section names on the bank frame will (try to) move all its items to the backpack.
Looking at this a bit more closely:
You don't need to build an item list. The section has the item list already.
Because this is multiple independent operations that affect client state, we need to move this loop into coros. PTAL at https://warcraft.wiki.gg/wiki/API_coroutine.create and associated functions and build a coroutine loop so the client doesn't lock up on large transfers.
Thanks for the feedback, @Cidan . I'll look into it and prepare an update.
awesome! later tonight, I'm going to add an async module to help, as this is a pattern I need in a few other places as well. this will make it relatively easy for you to implement this feature.
awesome! later tonight, I'm going to add an async module to help, as this is a pattern I need in a few other places as well. this will make it relatively easy for you to implement this feature.
Well, guess I'll wait for that module then!
@fiote I've just added a new module, Async, that will make this easier. You can now do the following:
async:Do(function()
for _, item in pairs(someContainer) do
C_Container.UseContainerItem(bagid, slotid)
async.Yield()
end
end)
The above function will process the loop one per frame, smoothing out the calls. You can also batch your execution if you want, i.e. keep a counter in the function and use some modulo to yield.
This is in main right now.
Okay. I think I did all that was requested.
I only checked the signing part after I was done so I'm not sure it's good enough. If you want me to, I can drop this PR, re-do all my changes (now signed) and PR again.
Also, not sure this is related to this PR. This feature works really well to (try to) send a whole category to the mail attachments. I already did it locally to test things and it's working great so far. Do you think it's part of the same feature (it's basically an extra if and that's it), or it's better as a new PR later on?
Also, not sure this is related to this PR. This feature works really well to (try to) send a whole category to the mail attachments. I already did it locally to test things and it's working great so far. Do you think it's part of the same feature (it's basically an extra if and that's it), or it's better as a new PR later on?
Edit: it's basically this, hah
Instead of checking here, look at hooks.lua, OpenInteractionWindow. We already detect if the mail event is triggered -- just create a new addon.atMailbox global and set it to true/false (false at the Close function under it).
Instead of checking here, look at hooks.lua, OpenInteractionWindow. We already detect if the mail event is triggered -- just create a new addon.atMailbox global and set it to true/false (false at the Close function under it).
That doesn't really work. That event (or the MAIL_SHOW, that I also checked) triggers when you open the inbox. But that's not enough. You need to check if the 'compose mail' frame is actually there, because since the 'move-to-other-container' method being used is the UseContainerItem, if you try to call that while on the inbox is open panel, it'll actually try to use/consume any item.
My way checks if the attachment frame is actually there/visible so the game will try to move items instead of using them when the UseContainerItem is called.
it'll actually try to use/consume any item - there is no destination for moving them.
Okay, this means we need to make a few changes:
Move the atMail boolean to a function, addon:AtSendMail. You can use init.lua to store this function.
At the time of click, you need to detect which vessel you are transferring to/from -- bank or mail, and store that.
On each click call, you need to make sure that the previous state of at bank or mail still applies, and the user didn't close the frame, otherwise items will be consumed mid-loop.
Do we want to support other interaction types? How about player trading?
Adding player trading on top of this sounds straightforward. I'll check this new approach and come back to you.
Now back to the previous comment (the bank-still-stacked one). The whole thing used that bagkind to work, because when generating the list I need to know where i'm searching. That said, now I know the section.content.cells.$.data have their bagid and slotid on it. Slotid is not really important for me, but bagid seems to be enough for me to detect if this click/section is on the bank or the backpack (https://wowwiki-archive.fandom.com/wiki/BagId)... I can use those constants to identity the 'bag kind'.
You want me to add that logic as a function/method somewhere ... ? or just have it inside my onTitleRightClick.
Oh also, let me think on bag kind for a bit. I really don't want to attach kind to a section if I can help it.
I kinda made all that I wanted to do, hah. You can check it if you want it, otherwise I'll wait for your updates to update my changes.
Merging conficts solved.
I am so sorry for the delay here -- I did not forget, but during testing a huge, massive set of bugs were discovered with how item updates work, and I had to rewrite a significant portion of the core addon systems.
Can you merge main, and try again on this?
I am so sorry for the delay here -- I did not forget, but during testing a huge, massive set of bugs were discovered with how item updates work, and I had to rewrite a significant portion of the core addon systems.
Can you merge main, and try again on this?
Sure, I'll take a look again!
Okay, looks like you changed a lot of things. So, question time! - I thought of trying re-doing things my way but you'll probably review it with "this can be done this other way, please adjust it" so I'm asking first and doing second, lol.
From local function onTitleRightClick(section)
if (kind == const.BAG_KIND.BACKPACK) then
source = items.itemsByBagAndSlot;
else
source = items.bankItemsByBagAndSlot;
end
items.itemsByBagAndSlot is gone. You still use "itemsByBagAndSlot" on your code, but its not where it was. How can I access it (the goal here is to get the inventory itemlist).
items.bankItemsByBagAndSlot is also gone. "bankItemsByBagAndSlot" variable name itself is also gone. How can I get the bank itemlist?
That code was actually unnecessary back then already.
Ok, I merged your changes and made it work again. I also expanded the feature to allow you to right-click the category header when a npcshop window is open to sell it. Since this one is stretching a bit the goal of moving items, i add a config entry for it, to players have to opt-in.
Can you merge in the latest main in and test this? I made a few changes.
@fiote When withdrawing from the bank or depositing to the bank, if there are virtual stacks of items in a category, this feature only moves the first item in a virtual stack and not all items in a virtual stack. i.e. buy 5 identical starter weapons, turn on stacking, and disable unstack at shop, then try to move things and it will only move one item at a time. Can you fix this?
|
2025-04-01T04:10:16.586309
| 2021-09-01T15:44:54
|
985312087
|
{
"authors": [
"Markl121",
"hud-code"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13786",
"repo": "CircleCI-Public/path-filtering-orb",
"url": "https://github.com/CircleCI-Public/path-filtering-orb/issues/19"
}
|
gharchive/issue
|
Include link to Dynamic Configuration documentation
Describe Request:
It would be great to get links to the dynamic config documentation for the path-filtering orb
Examples:
Dynamic Config: https://circleci.com/docs/2.0/dynamic-config/
Github documentation: https://github.com/CircleCI-Public/api-preview-docs/blob/master/docs/setup-workflows.md#concepts
Supporting Documentation Links:
Dynamic config docs link has been added to the orb docs as Homepage in https://github.com/CircleCI-Public/path-filtering-orb/pull/63
|
2025-04-01T04:10:16.594201
| 2015-12-24T09:33:33
|
123781311
|
{
"authors": [
"ChrisAubuchon",
"scalp42"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13787",
"repo": "CiscoCloud/mesos-consul",
"url": "https://github.com/CiscoCloud/mesos-consul/issues/54"
}
|
gharchive/issue
|
Stale services in Consul
After multiple deployments, I have a single service running in Marathon:
But unfortunately, Consul has some stale info in it:
The only app per Marathon should be running on 31263 but you can see that Consul is populated with some old tasks:
Any ideas ?
Thanks a lot in advance!
Check fix/service-cleanup branch. The wrong address was being used at deregistration
@ChrisAubuchon thanks a lot for the help!! Do you know if there'll be a master branch with both branches merged in ? (Trying to figure if we could do without forking)
There will be. Can you test the fix/service-cleanup branch and let me know if it fixes your issue? If it does I'll merge it in.
Do you know by any chance how to pull the branch with go get (without forking) ?
Going to fork for now to get this going.
I would run git clone -b fix/service-cleanup https://github.com/CiscoCloud/mesos-consul instead of go get
/opt/go/src/github.com/scalp42/mesos-consul/main.go:73: c.ServiceName undefined (type *config.Config has no field or method ServiceName)
/opt/go/src/github.com/scalp42/mesos-consul/main.go:74: c.ServiceTags undefined (type *config.Config has no field or method ServiceTags)
I merged it to master.
Looks good! Thanks @ChrisAubuchon
|
2025-04-01T04:10:16.634152
| 2023-01-24T20:43:23
|
1555639300
|
{
"authors": [
"fmunozmiranda",
"mschedrin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13788",
"repo": "CiscoISE/terraform-provider-ciscoise",
"url": "https://github.com/CiscoISE/terraform-provider-ciscoise/issues/77"
}
|
gharchive/issue
|
ConditionReference support in children section of resource ciscoise_network_access_authorization_rules
Prerequisites
[x] Have you tested the operation in the API directly?
[x] Do you have the latest Terraform provider version?
[x] Review the compatibility matrix before opening an issue.
Describe the bug
I attempt to create authorization rule and use condition_type = "ConditionReference" in children section of resource parameters. Here is resource configuration I am using:
resource "ciscoise_network_access_authorization_rules" "complex_condition" {
provider = ciscoise
depends_on = [ciscoise_authorization_profile.this]
parameters {
policy_id = ciscoise_network_access_policy_set.this["Wired_MAB_tf"].item.0.id
profile = ["APs_tf"]
rule {
condition {
condition_type = "ConditionAndBlock"
is_negate = false
children {
condition_type = "ConditionAttributes"
is_negate = false
dictionary_name = "EndPoints"
attribute_name = "LogicalProfile"
operator = "equals"
attribute_value = "APs"
}
children {
condition_type = "ConditionReference"
is_negate = false
id = "8bd7ac6e-0910-445e-9d7e-35327cc54e6b"
}
}
default = false
name = "rule with complex condition"
rank = 0
state = "enabled"
}
security_group = "Network_Services"
}
}
Terraform provider returns following error:
โ Error: Unsupported argument
โ
โ on network_access_authorization_rules.tf line 71, in resource "ciscoise_network_access_authorization_rules" "complex_condition":
โ 71: id = "8bd7ac6e-0910-445e-9d7e-35327cc54e6b"
โ
โ An argument named "id" is not expected here.
โต
ERRO[0001] Terraform invocation failed in /home/admmikshc/dev/terragrunt/tg-ise/.terragrunt-cache/_2EkuIjIiLp-E-_u9nri2yIGS7E/xQL-nJdlNnGL1jgOGimv-mCilCw/tf-ise prefix=[/home/admmikshc/dev/terragrunt/tg-ise]
ERRO[0001] 1 error occurred:
* exit status 1
Apparently parameter id is not supported in children section, however id is essential for creating condition of type ConditionReference. I have confirmed that it works with direct API requests.
Request:
{
"profile": [
"APs_tf"
],
"rule": {
"condition": {
"conditionType": "ConditionAndBlock",
"isNegate": false,
"children": [
{
"conditionType": "ConditionAttributes",
"isNegate": false,
"dictionaryName": "EndPoints",
"attributeName": "LogicalProfile",
"operator": "equals",
"attributeValue": "APs"
},
{
"conditionType": "ConditionReference",
"isNegate": false,
"id": "8bd7ac6e-0910-445e-9d7e-35327cc54e6b"
}
]
},
"default": false,
"name": "rule with complex condition4",
"rank": 0,
"state": "enabled"
},
"securityGroup": "Network_Services"
}
Response http code is 201 and payload is:
{
"version": "1.0.0",
"response": {
"rule": {
"default": false,
"id": "15ce7bc1-fe6d-45d0-b165-91a6039b9f0f",
"name": "rule with complex condition4",
"hitCounts": 0,
"rank": 0,
"state": "enabled",
"condition": {
"link": null,
"conditionType": "ConditionAndBlock",
"isNegate": false,
"children": [
{
"link": null,
"conditionType": "ConditionAttributes",
"isNegate": false,
"dictionaryName": "EndPoints",
"attributeName": "LogicalProfile",
"operator": "equals",
"dictionaryValue": null,
"attributeValue": "APs"
},
{
"link": null,
"conditionType": "ConditionReference",
"isNegate": false,
"name": "Wired_MAB",
"id": "8bd7ac6e-0910-445e-9d7e-35327cc54e6b",
"description": "A condition to match MAC Authentication Bypass service based authentication requests from switches, according to the corresponding MAB attributes defined in the device profile."
}
]
}
},
"profile": [
"APs_tf"
],
"securityGroup": "Network_Services",
"link": {
"rel": "self",
"href": "https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/15ce7bc1-fe6d-45d0-b165-91a6039b9f0f",
"type": "application/json"
}
}
}
Expected behavior
Parameter id as well as other parameters necessary for condition_type = "ConditionReference" should be supported in children section.
Environment (please complete the following information):
ISE version and patch: <IP_ADDRESS>8 Patch 3
Terraform version: 1.3.6
ISE provider version: 0.6.11-beta
OS version: Ubuntu 20.04
Hey @mschedrin , there's a new version (0.6.12-beta) of provider, please try it again and let us know if it works now.
It works now ๐
I was too quick to confirm that resource works as expected. It turns out that terraform state is not handled properly with a resource created with type "ConditionReference" in children section. Terraform now suggests updating the resource in place upon every plan/update action and fails doing that. Based on my understanding the provider fails to match values in API reply with tfstate, additionally it fails to generate correct payload when updating the resource. Parameters in tfstate seem to be correct.
Terraform output when running terraform apply:
Terraform will perform the following actions:
# ciscoise_network_access_authorization_rules.complex_condition will be updated in-place
~ resource "ciscoise_network_access_authorization_rules" "complex_condition" {
id = "id:=8079a15a-a14f-47bd-8d8f-37a6710d464f\\name:=rule with complex condition\\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a"
# (1 unchanged attribute hidden)
~ parameters {
id = "8079a15a-a14f-47bd-8d8f-37a6710d464f"
# (4 unchanged attributes hidden)
~ rule {
id = "8079a15a-a14f-47bd-8d8f-37a6710d464f"
name = "rule with complex condition"
# (4 unchanged attributes hidden)
~ condition {
# (5 unchanged attributes hidden)
~ children {
+ attribute_name = "LogicalProfile"
+ attribute_value = "APs"
+ dictionary_name = "EndPoints"
+ operator = "equals"
# (3 unchanged attributes hidden)
}
~ children {
+ id = "8bd7ac6e-0910-445e-9d7e-35327cc54e6b"
# (3 unchanged attributes hidden)
}
}
}
}
}
Plan: 0 to add, 1 to change, 0 to destroy.
ciscoise_network_access_authorization_rules.complex_condition: Modifying... [id=id:=8079a15a-a14f-47bd-8d8f-37a6710d464f\name:=rule with complex condition\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
โท
โ Error: Failure when executing UpdateNetworkAccessAuthorizationRuleByID
โ
โ with ciscoise_network_access_authorization_rules.complex_condition,
โ on network_access_authorization_rules.tf line 45, in resource "ciscoise_network_access_authorization_rules" "complex_condition":
โ 45: resource "ciscoise_network_access_authorization_rules" "complex_condition" {
โ
โ error with operation UpdateNetworkAccessAuthorizationRuleById
โ {
โ "message" : "request has bad input format in the body",
โ "code" : 400
โ }
โต
Terraform state of corresponding resource:
> terraform state show ciscoise_network_access_authorization_rules.complex_condition
# ciscoise_network_access_authorization_rules.complex_condition:
resource "ciscoise_network_access_authorization_rules" "complex_condition" {
id = "id:=8079a15a-a14f-47bd-8d8f-37a6710d464f\\name:=rule with complex condition\\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a"
item = [
{
link = [
{
href = "https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f"
rel = "self"
type = "application/json"
},
]
profile = [
"APs_tf",
]
rule = [
{
condition = [
{
attribute_name = ""
attribute_value = ""
children = [
{
condition_type = "ConditionAttributes"
id = ""
is_negate = "false"
link = []
},
{
condition_type = "ConditionReference"
id = ""
is_negate = "false"
link = []
},
]
condition_type = "ConditionAndBlock"
dates_range = []
dates_range_exception = []
description = ""
dictionary_name = ""
dictionary_value = ""
hours_range = []
hours_range_exception = []
id = ""
is_negate = "false"
link = []
name = ""
operator = ""
week_days = []
week_days_exception = []
},
]
default = "false"
hit_counts = 0
id = "8079a15a-a14f-47bd-8d8f-37a6710d464f"
name = "rule with complex condition"
rank = 1
state = "enabled"
},
]
security_group = "Network_Services"
},
]
parameters {
id = "8079a15a-a14f-47bd-8d8f-37a6710d464f"
link = [
{
href = "https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f"
rel = "self"
type = "application/json"
},
]
policy_id = "ac8db424-5856-468f-a93e-b913d6e8d57a"
profile = [
"APs_tf",
]
security_group = "Network_Services"
rule {
default = "false"
hit_counts = 0
id = "8079a15a-a14f-47bd-8d8f-37a6710d464f"
name = "rule with complex condition"
rank = 1
state = "enabled"
condition {
condition_type = "ConditionAndBlock"
is_negate = "false"
link = []
week_days = []
week_days_exception = []
children {
attribute_name = "LogicalProfile"
attribute_value = "APs"
condition_type = "ConditionAttributes"
dictionary_name = "EndPoints"
is_negate = "false"
link = []
operator = "equals"
}
children {
condition_type = "ConditionReference"
id = "8bd7ac6e-0910-445e-9d7e-35327cc54e6b"
is_negate = "false"
link = []
}
}
}
}
}
Relevant terraaform debug log:
==============================================================================
~~~ REQUEST ~~~
GET /api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f HTTP/1.1
HOST : <IP_ADDRESS>:443
HEADERS:
Accept: application/json
Authorization: Basic YWRtaW46TTBuZXQxMjM=
Content-Type: application/json
User-Agent: go-resty/2.7.0 (https://github.com/go-resty/resty)
BODY :
***** NO CONTENT *****
------------------------------------------------------------------------------
~~~ RESPONSE ~~~
STATUS : 200
PROTO : HTTP/1.1
RECEIVED AT : 2023-01-31T09:39:06.293004224-08:00
TIME DURATION: 253.939723ms
HEADERS :
Cache-Control: no-cache, no-store, must-revalidate
Connection: keep-alive
Content-Security-Policy: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
Content-Type: application/json
Date: Tue, 31 Jan 2023 17:38:35 GMT
Expires: Thu, 01 Jan 1970 00:00:00 GMT
Pragma: no-cache
Server:
Strict-Transport-Security: max-age=31536000; includeSubDomains
Vary: accept-encoding
X-Content-Security-Policy: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
X-Content-Type-Options: nosniff
X-Frame-Options: SAMEORIGIN
X-Request-Id: 16c5c080-a18e-11ed-8b18-424b604f84ca
X-Webkit-Csp: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
X-Xss-Protection: 1; mode=block
BODY :
{
"version": "1.0.0",
"response": {
"rule": {
"default": false,
"id": "8079a15a-a14f-47bd-8d8f-37a6710d464f",
"name": "rule with complex condition",
"hitCounts": 0,
"rank": 1,
"state": "enabled",
"condition": {
"link": null,
"conditionType": "ConditionAndBlock",
"isNegate": false,
"children": [
{
"link": null,
"conditionType": "ConditionAttributes",
"isNegate": false,
"dictionaryName": "EndPoints",
"attributeName": "LogicalProfile",
"operator": "equals",
"dictionaryValue": null,
"attributeValue": "APs"
},
{
"link": null,
"conditionType": "ConditionReference",
"isNegate": false,
"name": "Wired_MAB",
"id": "8bd7ac6e-0910-445e-9d7e-35327cc54e6b",
"description": "A condition to match MAC Authentication Bypass service based authentication requests from switches, according to the corresponding MAB attributes defined in the device profile."
}
]
}
},
"profile": [
"APs_tf"
],
"securityGroup": "Network_Services",
"link": {
"rel": "self",
"href": "https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f",
"type": "application/json"
}
}
}
==============================================================================: timestamp=2023-01-31T09:39:06.293-0800
2023-01-31T09:39:06.293-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] Retrieved response {"response":{"link":{"href":"https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f","rel":"self","type":"application/json"},"profile":["APs_tf"],"rule":{"condition":{"conditionType":"ConditionAndBlock","isNegate":false,"children":[{"conditionType":"ConditionAttributes","isNegate":false,"dictionaryName":"EndPoints","attributeName":"LogicalProfile","operator":"equals","attributeValue":"APs"},{"conditionType":"ConditionReference","isNegate":false}]},"default":false,"hitCounts":0,"id":"8079a15a-a14f-47bd-8d8f-37a6710d464f","name":"rule with complex condition","rank":1,"state":"enabled"},"securityGroup":"Network_Services"},"version":"1.0.0"}: timestamp=2023-01-31T09:39:06.293-0800
2023-01-31T09:39:06.298-0800 [WARN] Provider "registry.terraform.io/ciscoise/ciscoise" produced an unexpected new value for ciscoise_network_access_authorization_rules.complex_condition during refresh.
- .parameters[0].rule[0].condition[0].children[0].attribute_name: was cty.StringVal("LogicalProfile"), but now cty.StringVal("")
- .parameters[0].rule[0].condition[0].children[0].attribute_value: was cty.StringVal("APs"), but now cty.StringVal("")
- .parameters[0].rule[0].condition[0].children[0].dictionary_name: was cty.StringVal("EndPoints"), but now cty.StringVal("")
- .parameters[0].rule[0].condition[0].children[0].operator: was cty.StringVal("equals"), but now cty.StringVal("")
- .parameters[0].rule[0].condition[0].children[1].id: was cty.StringVal("8bd7ac6e-0910-445e-9d7e-35327cc54e6b"), but now cty.StringVal("")
2023-01-31T09:39:06.316-0800 [WARN] Provider "registry.terraform.io/ciscoise/ciscoise" produced an invalid plan for ciscoise_network_access_authorization_rules.complex_condition, but we are tolerating it because it is using the legacy plugin SDK.
The following problems may be the cause of any confusing errors from downstream operations:
- .parameters[0].id: planned value cty.StringVal("8079a15a-a14f-47bd-8d8f-37a6710d464f") for a non-computed attribute
2023-01-31T09:39:06.318-0800 [DEBUG] provider.stdio: received EOF, stopping recv loop: err="rpc error: code = Unavailable desc = error reading from server: EOF"
2023-01-31T09:39:06.324-0800 [DEBUG] provider: plugin process exited: path=.terraform/providers/registry.terraform.io/ciscoise/ciscoise/0.6.12-beta/linux_amd64/terraform-provider-ciscoise_v0.6.12-beta pid=558022
2023-01-31T09:39:06.324-0800 [DEBUG] provider: plugin exited
2023-01-31T09:39:06.334-0800 [DEBUG] building apply graph to check for errors
2023-01-31T09:39:06.334-0800 [DEBUG] adding implicit provider configuration provider["terraform.io/builtin/terraform"], implied first by data.terraform_remote_state.dnac (expand)
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "ciscoise_authorization_profile.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_conditions.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "data.ciscoise_network_access_conditions_for_authentication_rule.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "data.ciscoise_network_access_conditions_for_policy_set.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authorization_rules.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "data.ciscoise_network_access_conditions_for_authorization_rule.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authentication_rules.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_policy_set.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "data.terraform_remote_state.dnac (expand)" (*terraform.nodeExpandApplyableResource) needs provider["terraform.io/builtin/terraform"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authorization_rules.complex_condition (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.334-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authorization_rules.complex_condition" (*terraform.NodeApplyableResourceInstance) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.335-0800 [ERROR] AttachSchemaTransformer: No provider config schema available for provider["terraform.io/builtin/terraform"]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_policy_set.this (expand)" references: [var.policy_sets]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "data.terraform_remote_state.dnac (expand)" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "data.ciscoise_network_access_conditions_for_policy_set.this (expand)" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "var.authz_profiles" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "var.authc_rules" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "local.ciscoise_network_access_conditions_for_policy_set (expand)" references: [data.ciscoise_network_access_conditions_for_policy_set.this (expand)]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "provider[\"terraform.io/builtin/terraform\"]" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authentication_rules.this (expand)" references: [var.authc_rules]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "var.authz_rules" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authorization_rules.complex_condition (expand)" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "data.ciscoise_network_access_conditions_for_authorization_rule.this (expand)" references: []
2023-01-31T09:39:06.335-0800 [INFO] ReferenceTransformer: reference not found: "ciscoise_authorization_profile.this#destroy"
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authorization_rules.complex_condition" references: [ciscoise_authorization_profile.this (expand) ciscoise_network_access_policy_set.this (expand)]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authorization_rules.this (expand)" references: [var.authz_rules]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_conditions.this (expand)" references: [var.conditions]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "data.ciscoise_network_access_conditions_for_authentication_rule.this (expand)" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "var.conditions" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "var.policy_sets" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "local.network_access_conditions_for_authc_rule (expand)" references: [data.ciscoise_network_access_conditions_for_authentication_rule.this (expand)]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "local.network_access_conditions_for_authz_rule (expand)" references: [data.ciscoise_network_access_conditions_for_authorization_rule.this (expand)]
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "provider[\"registry.terraform.io/ciscoise/ciscoise\"]" references: []
2023-01-31T09:39:06.335-0800 [DEBUG] ReferenceTransformer: "ciscoise_authorization_profile.this (expand)" references: [var.authz_profiles]
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: data.terraform_remote_state.dnac (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: local.ciscoise_network_access_conditions_for_policy_set (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: provider["terraform.io/builtin/terraform"] is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: ciscoise_network_access_authentication_rules.this (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: ciscoise_network_access_authorization_rules.this (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: ciscoise_network_access_conditions.this (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: data.ciscoise_network_access_conditions_for_policy_set.this (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: local.network_access_conditions_for_authz_rule (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: local.network_access_conditions_for_authc_rule (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: data.ciscoise_network_access_conditions_for_authorization_rule.this (expand) is no longer needed, removing
2023-01-31T09:39:06.336-0800 [DEBUG] pruneUnusedNodes: data.ciscoise_network_access_conditions_for_authentication_rule.this (expand) is no longer needed, removing
2023-01-31T09:39:06.341-0800 [INFO] backend/local: apply calling Apply
2023-01-31T09:39:06.341-0800 [DEBUG] Building and walking apply graph for NormalMode plan
2023-01-31T09:39:06.341-0800 [DEBUG] adding implicit provider configuration provider["terraform.io/builtin/terraform"], implied first by data.terraform_remote_state.dnac (expand)
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authentication_rules.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authorization_rules.complex_condition (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "data.ciscoise_network_access_conditions_for_authentication_rule.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "data.terraform_remote_state.dnac (expand)" (*terraform.nodeExpandApplyableResource) needs provider["terraform.io/builtin/terraform"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_policy_set.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "ciscoise_authorization_profile.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authorization_rules.complex_condition" (*terraform.NodeApplyableResourceInstance) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_conditions.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "data.ciscoise_network_access_conditions_for_policy_set.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "data.ciscoise_network_access_conditions_for_authorization_rule.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [DEBUG] ProviderTransformer: "ciscoise_network_access_authorization_rules.this (expand)" (*terraform.nodeExpandApplyableResource) needs provider["registry.terraform.io/ciscoise/ciscoise"]
2023-01-31T09:39:06.342-0800 [ERROR] AttachSchemaTransformer: No provider config schema available for provider["terraform.io/builtin/terraform"]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "data.terraform_remote_state.dnac (expand)" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "var.policy_sets" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "local.network_access_conditions_for_authc_rule (expand)" references: [data.ciscoise_network_access_conditions_for_authentication_rule.this (expand)]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "provider[\"terraform.io/builtin/terraform\"]" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authorization_rules.this (expand)" references: [var.authz_rules]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_policy_set.this (expand)" references: [var.policy_sets]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "ciscoise_authorization_profile.this (expand)" references: [var.authz_profiles]
2023-01-31T09:39:06.342-0800 [INFO] ReferenceTransformer: reference not found: "ciscoise_authorization_profile.this#destroy"
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authorization_rules.complex_condition" references: [ciscoise_authorization_profile.this (expand) ciscoise_network_access_policy_set.this (expand)]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authentication_rules.this (expand)" references: [var.authc_rules]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_conditions.this (expand)" references: [var.conditions]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "data.ciscoise_network_access_conditions_for_authentication_rule.this (expand)" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "data.ciscoise_network_access_conditions_for_policy_set.this (expand)" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "var.authc_rules" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "var.conditions" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "local.ciscoise_network_access_conditions_for_policy_set (expand)" references: [data.ciscoise_network_access_conditions_for_policy_set.this (expand)]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "ciscoise_network_access_authorization_rules.complex_condition (expand)" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "var.authz_rules" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "var.authz_profiles" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "local.network_access_conditions_for_authz_rule (expand)" references: [data.ciscoise_network_access_conditions_for_authorization_rule.this (expand)]
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "provider[\"registry.terraform.io/ciscoise/ciscoise\"]" references: []
2023-01-31T09:39:06.342-0800 [DEBUG] ReferenceTransformer: "data.ciscoise_network_access_conditions_for_authorization_rule.this (expand)" references: []
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: ciscoise_network_access_authentication_rules.this (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: ciscoise_network_access_conditions.this (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: local.ciscoise_network_access_conditions_for_policy_set (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: local.network_access_conditions_for_authz_rule (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: data.ciscoise_network_access_conditions_for_authorization_rule.this (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: ciscoise_network_access_authorization_rules.this (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: data.ciscoise_network_access_conditions_for_policy_set.this (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: local.network_access_conditions_for_authc_rule (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: data.terraform_remote_state.dnac (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: provider["terraform.io/builtin/terraform"] is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] pruneUnusedNodes: data.ciscoise_network_access_conditions_for_authentication_rule.this (expand) is no longer needed, removing
2023-01-31T09:39:06.346-0800 [DEBUG] Starting graph walk: walkApply
2023-01-31T09:39:06.346-0800 [DEBUG] created provider logger: level=debug
2023-01-31T09:39:06.347-0800 [INFO] provider: configuring client automatic mTLS
2023-01-31T09:39:06.354-0800 [DEBUG] provider: starting plugin: path=.terraform/providers/registry.terraform.io/ciscoise/ciscoise/0.6.12-beta/linux_amd64/terraform-provider-ciscoise_v0.6.12-beta args=[.terraform/providers/registry.terraform.io/ciscoise/ciscoise/0.6.12-beta/linux_amd64/terraform-provider-ciscoise_v0.6.12-beta]
2023-01-31T09:39:06.354-0800 [DEBUG] provider: plugin started: path=.terraform/providers/registry.terraform.io/ciscoise/ciscoise/0.6.12-beta/linux_amd64/terraform-provider-ciscoise_v0.6.12-beta pid=558032
2023-01-31T09:39:06.354-0800 [DEBUG] provider: waiting for RPC address: path=.terraform/providers/registry.terraform.io/ciscoise/ciscoise/0.6.12-beta/linux_amd64/terraform-provider-ciscoise_v0.6.12-beta
2023-01-31T09:39:06.360-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: configuring server automatic mTLS: timestamp=2023-01-31T09:39:06.360-0800
2023-01-31T09:39:06.381-0800 [DEBUG] provider.terraform-provider-ciscoise_v0.6.12-beta: plugin address: network=unix address=/tmp/plugin1800883229 timestamp=2023-01-31T09:39:06.381-0800
2023-01-31T09:39:06.381-0800 [DEBUG] provider: using plugin: version=5
2023-01-31T09:39:06.545-0800 [WARN] ValidateProviderConfig from "provider[\"registry.terraform.io/ciscoise/ciscoise\"]" changed the config value, but that value is unused
2023-01-31T09:39:06.565-0800 [WARN] Provider "registry.terraform.io/ciscoise/ciscoise" produced an invalid plan for ciscoise_network_access_authorization_rules.complex_condition, but we are tolerating it because it is using the legacy plugin SDK.
The following problems may be the cause of any confusing errors from downstream operations:
- .parameters[0].id: planned value cty.StringVal("8079a15a-a14f-47bd-8d8f-37a6710d464f") for a non-computed attribute
2023-01-31T09:39:06.566-0800 [INFO] Starting apply for ciscoise_network_access_authorization_rules.complex_condition
2023-01-31T09:39:06.567-0800 [DEBUG] ciscoise_network_access_authorization_rules.complex_condition: applying the planned Update change
2023-01-31T09:39:06.581-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] Beginning NetworkAccessAuthorizationRules update for id=[id:=8079a15a-a14f-47bd-8d8f-37a6710d464f\name:=rule with complex condition\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]: timestamp=2023-01-31T09:39:06.581-0800
2023-01-31T09:39:06.581-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] Selecting method. Method 1 [true]: timestamp=2023-01-31T09:39:06.581-0800
2023-01-31T09:39:06.581-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] Selecting method. Method 2 [true]: timestamp=2023-01-31T09:39:06.581-0800
2023-01-31T09:39:06.586-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] ID used for update operation 8079a15a-a14f-47bd-8d8f-37a6710d464f: timestamp=2023-01-31T09:39:06.586-0800
2023-01-31T09:39:06.587-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] request sent => {"link":{"href":"https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f","rel":"self","type":"application/json"},"profile":["APs_tf"],"rule":{"condition":{"conditionType":"ConditionAndBlock","isNegate":false,"children":[{"conditionType":"ConditionAttributes","isNegate":false,"dictionaryName":"EndPoints","attributeName":"LogicalProfile","operator":"equals","attributeValue":"APs"},{"conditionType":"ConditionReference","isNegate":false}]},"default":false,"hitCounts":0,"id":"8079a15a-a14f-47bd-8d8f-37a6710d464f","name":"rule with complex condition","rank":1,"state":"enabled"},"securityGroup":"Network_Services"}: timestamp=2023-01-31T09:39:06.587-0800
2023-01-31T09:39:06.686-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] [RESTY]
==============================================================================
~~~ REQUEST ~~~
PUT /api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f HTTP/1.1
HOST : <IP_ADDRESS>:443
HEADERS:
Accept: application/json
Authorization: Basic YWRtaW46TTBuZXQxMjM=
Content-Type: application/json
User-Agent: go-resty/2.7.0 (https://github.com/go-resty/resty)
BODY :
{
"link": {
"href": "https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/ac8db424-5856-468f-a93e-b913d6e8d57a/authorization/8079a15a-a14f-47bd-8d8f-37a6710d464f",
"rel": "self",
"type": "application/json"
},
"profile": [
"APs_tf"
],
"rule": {
"condition": {
"conditionType": "ConditionAndBlock",
"isNegate": false,
"children": [
{
"conditionType": "ConditionAttributes",
"isNegate": false,
"dictionaryName": "EndPoints",
"attributeName": "LogicalProfile",
"operator": "equals",
"attributeValue": "APs"
},
{
"conditionType": "ConditionReference",
"isNegate": false
}
]
},
"default": false,
"hitCounts": 0,
"id": "8079a15a-a14f-47bd-8d8f-37a6710d464f",
"name": "rule with complex condition",
"rank": 1,
"state": "enabled"
},
"securityGroup": "Network_Services"
}
------------------------------------------------------------------------------
~~~ RESPONSE ~~~
STATUS : 400
PROTO : HTTP/1.1
RECEIVED AT : 2023-01-31T09:39:06.685735419-08:00
TIME DURATION: 98.219384ms
HEADERS :
Cache-Control: no-cache, no-store, must-revalidate
Connection: keep-alive
Content-Security-Policy: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
Content-Type: application/json
Date: Tue, 31 Jan 2023 17:38:35 GMT
Expires: Thu, 01 Jan 1970 00:00:00 GMT
Pragma: no-cache
Server:
Set-Cookie: JSESSIONIDSSO=BDD7A1F4C2DCDE643AEFA39E281F5612; Path=/; Secure; HttpOnly, APPSESSIONID=9DD2D6A7559A72224868F8D358EE94AD; Path=/api; Secure; HttpOnly
Strict-Transport-Security: max-age=31536000; includeSubDomains
Vary: accept-encoding
X-Content-Security-Policy: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
X-Content-Type-Options: nosniff
X-Frame-Options: SAMEORIGIN
X-Request-Id: 17278f90-a18e-11ed-8b18-424b604f84ca
X-Webkit-Csp: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
X-Xss-Protection: 1; mode=block
BODY :
{
"message": "request has bad input format in the body",
"code": 400
}
==============================================================================: timestamp=2023-01-31T09:39:06.685-0800
2023-01-31T09:39:06.686-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.12-beta: 2023/01/31 09:39:06 [DEBUG] resty response for update operation => {
"message" : "request has bad input format in the body",
"code" : 400
}: timestamp=2023-01-31T09:39:06.685-0800
2023-01-31T09:39:06.704-0800 [ERROR] vertex "ciscoise_network_access_authorization_rules.complex_condition" error: Failure when executing UpdateNetworkAccessAuthorizationRuleByID
2023-01-31T09:39:06.709-0800 [DEBUG] provider.stdio: received EOF, stopping recv loop: err="rpc error: code = Unavailable desc = error reading from server: EOF"
2023-01-31T09:39:06.713-0800 [DEBUG] provider: plugin process exited: path=.terraform/providers/registry.terraform.io/ciscoise/ciscoise/0.6.12-beta/linux_amd64/terraform-provider-ciscoise_v0.6.12-beta pid=558032
2023-01-31T09:39:06.713-0800 [DEBUG] provider: plugin exited
Hey @mschedrin, could you pass me a valid request_body for the update?
@mschedrin thanks for report, try it again with new provider version v.0.6.13-beta and update us.
Tested 0.6.13. Resource in-place update now works and generates valid payload, however state handling still does not work correct. After terraform successfully creates ciscoise_network_access_authorization_rules resource it suggests doing in-place update for that resource each time I run apply/plan action.
Terraform output:
> terraform apply -auto-approve
data.terraform_remote_state.dnac: Reading...
data.terraform_remote_state.dnac: Read complete after 0s
ciscoise_authorization_profile.this["APs_tf"]: Refreshing state... [id=id:=55a20160-9b21-11ed-8b18-424b604f84ca\name:=APs_tf]
ciscoise_network_access_conditions.this["Guests"]: Refreshing state... [id=id:=e35a92ba-0a7b-462d-8832-b5bb36a87edb\name:=Guests]
data.ciscoise_network_access_conditions_for_authentication_rule.this: Reading...
data.ciscoise_network_access_conditions_for_authorization_rule.this: Reading...
data.ciscoise_network_access_conditions_for_policy_set.this: Reading...
data.ciscoise_network_access_conditions_for_authentication_rule.this: Read complete after 1s [id=1675340771]
data.ciscoise_network_access_conditions_for_authorization_rule.this: Read complete after 1s [id=1675340771]
data.ciscoise_network_access_conditions_for_policy_set.this: Read complete after 1s [id=1675340771]
ciscoise_network_access_policy_set.this["Wired_MAB_tf"]: Refreshing state... [id=id:=ac8db424-5856-468f-a93e-b913d6e8d57a\name:=Wired_MAB_tf]
ciscoise_network_access_authorization_rules.complex_condition: Refreshing state... [id=id:=de60e4f5-c48c-4e06-bfd2-ac0ad304d82f\name:=rule with complex condition\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
ciscoise_network_access_authentication_rules.this["MAB"]: Refreshing state... [id=id:=1d2c7e7f-42e8-44ba-b052-b7321902c8af\name:=MAB\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
ciscoise_network_access_authorization_rules.this["APs"]: Refreshing state... [id=id:=eb65aadb-2372-4ebe-979a-3bf6c64c6378\name:=APs\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
Terraform used the selected providers to generate the following execution plan. Resource actions are indicated with the following symbols:
~ update in-place
Terraform will perform the following actions:
# ciscoise_network_access_authorization_rules.complex_condition will be updated in-place
~ resource "ciscoise_network_access_authorization_rules" "complex_condition" {
id = "id:=de60e4f5-c48c-4e06-bfd2-ac0ad304d82f\\name:=rule with complex condition\\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a"
# (1 unchanged attribute hidden)
~ parameters {
id = "de60e4f5-c48c-4e06-bfd2-ac0ad304d82f"
# (4 unchanged attributes hidden)
~ rule {
id = "de60e4f5-c48c-4e06-bfd2-ac0ad304d82f"
name = "rule with complex condition"
# (4 unchanged attributes hidden)
~ condition {
# (5 unchanged attributes hidden)
~ children {
+ attribute_name = "LogicalProfile"
+ attribute_value = "APs"
+ dictionary_name = "EndPoints"
+ operator = "equals"
# (3 unchanged attributes hidden)
}
~ children {
+ id = "8bd7ac6e-0910-445e-9d7e-35327cc54e6b"
# (3 unchanged attributes hidden)
}
}
}
}
}
# ciscoise_network_access_authorization_rules.this["APs"] will be updated in-place
~ resource "ciscoise_network_access_authorization_rules" "this" {
id = "id:=eb65aadb-2372-4ebe-979a-3bf6c64c6378\\name:=APs\\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a"
# (2 unchanged attributes hidden)
~ parameters {
id = "eb65aadb-2372-4ebe-979a-3bf6c64c6378"
# (4 unchanged attributes hidden)
~ rule {
id = "eb65aadb-2372-4ebe-979a-3bf6c64c6378"
name = "APs"
~ rank = 1 -> 0
# (3 unchanged attributes hidden)
# (1 unchanged block hidden)
}
}
}
Plan: 0 to add, 2 to change, 0 to destroy.
ciscoise_network_access_authorization_rules.this["APs"]: Modifying... [id=id:=eb65aadb-2372-4ebe-979a-3bf6c64c6378\name:=APs\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
ciscoise_network_access_authorization_rules.complex_condition: Modifying... [id=id:=de60e4f5-c48c-4e06-bfd2-ac0ad304d82f\name:=rule with complex condition\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
ciscoise_network_access_authorization_rules.this["APs"]: Modifications complete after 1s [id=id:=eb65aadb-2372-4ebe-979a-3bf6c64c6378\name:=APs\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
ciscoise_network_access_authorization_rules.complex_condition: Modifications complete after 1s [id=id:=de60e4f5-c48c-4e06-bfd2-ac0ad304d82f\name:=rule with complex condition\policy_id:=ac8db424-5856-468f-a93e-b913d6e8d57a]
Apply complete! Resources: 0 added, 2 changed, 0 destroyed.
Terraform log where I first succesfully create the resource and then run terraform apply a couple times more and terraform does in-place update each time:
ise.log
@fmunozmiranda, after installing 0.6.14-beta resource ciscoise_network_access_authentication_rules stopped working completely. Here is what happens when I run terraform apply:
ciscoise_network_access_authorization_rules.complex_condition: Creating...
ciscoise_network_access_authentication_rules.this["MAB"]: Creation complete after 0s [id=id:=52c12fa9-83eb-47d0-8820-78b3663ab960\name:=MAB\policy_id:=25fda148-ce05-434a-8894-cd94e52875ec]
โท
โ Error: Failure when setting GetNetworkAccessAuthorizationRuleByID response
โ
โ with ciscoise_network_access_authorization_rules.complex_condition,
โ on network_access_authorization_rules.tf line 45, in resource "ciscoise_network_access_authorization_rules" "complex_condition":
โ 45: resource "ciscoise_network_access_authorization_rules" "complex_condition" {
โ
โ item.0.rule.0.condition.0.children.0.is_negate: '' expected type 'string', got unconvertible type 'bool', value: '0xc00003ef50'
โต
Relevant part of provider debug:
==============================================================================: timestamp=2023-02-13T04:56:22.158-0800
2023-02-13T04:56:22.159-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [DEBUG] Beginning NetworkAccessAuthorizationRules read for id=[id:=36b76f9a-9bb4-42b1-a4ec-2d02c769bd32\name:=rule with complex condition\policy_id:=25fda148-ce05-434a-8894-cd94e52875ec]: timestamp=2023-02-13T04:56:22.158-0800
2023-02-13T04:56:22.159-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [DEBUG] Selecting method. Method 1 [true]: timestamp=2023-02-13T04:56:22.159-0800
2023-02-13T04:56:22.159-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [DEBUG] Selecting method. Method 2 [true]: timestamp=2023-02-13T04:56:22.159-0800
2023-02-13T04:56:22.159-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [DEBUG] Selected method: GetNetworkAccessAuthorizationRuleByID: timestamp=2023-02-13T04:56:22.159-0800
2023-02-13T04:56:22.211-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [DEBUG] [RESTY]
==============================================================================
~~~ REQUEST ~~~
GET /api/v1/policy/network-access/policy-set/25fda148-ce05-434a-8894-cd94e52875ec/authorization/36b76f9a-9bb4-42b1-a4ec-2d02c769bd32 HTTP/1.1
HOST : <IP_ADDRESS>:443
HEADERS:
Accept: application/json
Authorization: Basic YWRtaW46TTBuZXQxMjM=
Content-Type: application/json
User-Agent: go-resty/2.7.0 (https://github.com/go-resty/resty)
BODY :
***** NO CONTENT *****
------------------------------------------------------------------------------
~~~ RESPONSE ~~~
STATUS : 200
PROTO : HTTP/1.1
RECEIVED AT : 2023-02-13T04:56:22.210638386-08:00
TIME DURATION: 51.569445ms
HEADERS :
Cache-Control: no-cache, no-store, must-revalidate
Connection: keep-alive
Content-Security-Policy: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
Content-Type: application/json
Date: Mon, 13 Feb 2023 12:56:15 GMT
Expires: Thu, 01 Jan 1970 00:00:00 GMT
Pragma: no-cache
Server:
Strict-Transport-Security: max-age=31536000; includeSubDomains
Vary: accept-encoding
X-Content-Security-Policy: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
X-Content-Type-Options: nosniff
X-Frame-Options: SAMEORIGIN
X-Request-Id: cd6077e0-ab9d-11ed-8b18-424b604f84ca
X-Webkit-Csp: default-src 'self'; script-src 'self' 'unsafe-inline' 'unsafe-eval'; style-src 'self' 'unsafe-inline'; img-src 'self' data:;
X-Xss-Protection: 1; mode=block
BODY :
{
"version": "1.0.0",
"response": {
"rule": {
"default": false,
"id": "36b76f9a-9bb4-42b1-a4ec-2d02c769bd32",
"name": "rule with complex condition",
"hitCounts": 0,
"rank": 0,
"state": "enabled",
"condition": {
"link": null,
"conditionType": "ConditionAndBlock",
"isNegate": false,
"children": [
{
"link": null,
"conditionType": "ConditionAttributes",
"isNegate": false,
"dictionaryName": "EndPoints",
"attributeName": "LogicalProfile",
"operator": "equals",
"dictionaryValue": null,
"attributeValue": "APs"
},
{
"link": null,
"conditionType": "ConditionReference",
"isNegate": false,
"name": "Wired_MAB",
"id": "8bd7ac6e-0910-445e-9d7e-35327cc54e6b",
"description": "A condition to match MAC Authentication Bypass service based authentication requests from switches, according to the corresponding MAB attributes defined in the device profile."
}
]
}
},
"profile": [
"APs_tf"
],
"securityGroup": "Network_Services",
"link": {
"rel": "self",
"href": "https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/25fda148-ce05-434a-8894-cd94e52875ec/authorization/36b76f9a-9bb4-42b1-a4ec-2d02c769bd32",
"type": "application/json"
}
}
}
==============================================================================: timestamp=2023-02-13T04:56:22.210-0800
2023-02-13T04:56:22.211-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [DEBUG] Retrieved response {"response":{"link":{"href":"https://<IP_ADDRESS>/api/v1/policy/network-access/policy-set/25fda148-ce05-434a-8894-cd94e52875ec/authorization/36b76f9a-9bb4-42b1-a4ec-2d02c769bd32","rel":"self","type":"application/json"},"profile":["APs_tf"],"rule":{"condition":{"conditionType":"ConditionAndBlock","isNegate":false,"children":[{"conditionType":"ConditionAttributes","isNegate":false,"dictionaryName":"EndPoints","attributeName":"LogicalProfile","operator":"equals","attributeValue":"APs"},{"conditionType":"ConditionReference","isNegate":false,"id":"8bd7ac6e-0910-445e-9d7e-35327cc54e6b"}]},"default":false,"hitCounts":0,"id":"36b76f9a-9bb4-42b1-a4ec-2d02c769bd32","name":"rule with complex condition","rank":0,"state":"enabled"},"securityGroup":"Network_Services"},"version":"1.0.0"}: timestamp=2023-02-13T04:56:22.211-0800
2023-02-13T04:56:22.211-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [ERROR] setting state: item.0.rule.0.condition.0.children.0.is_negate: '' expected type 'string', got unconvertible type 'bool', value: '0xc00003ef50': timestamp=2023-02-13T04:56:22.211-0800
2023-02-13T04:56:22.220-0800 [ERROR] vertex "ciscoise_network_access_authorization_rules.complex_condition" error: Failure when setting GetNetworkAccessAuthorizationRuleByID response
2023-02-13T04:56:22.403-0800 [INFO] provider.terraform-provider-ciscoise_v0.6.14-beta: 2023/02/13 04:56:22 [DEBUG] [RESTY]
==============================================================================
It works as expected in 0.6.15. Thank you.
|
2025-04-01T04:10:16.677732
| 2022-09-30T23:55:00
|
1393126501
|
{
"authors": [
"haileyhoyat",
"jpt-c"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13789",
"repo": "City-Bureau/city-scrapers-fresno",
"url": "https://github.com/City-Bureau/city-scrapers-fresno/pull/106"
}
|
gharchive/pull-request
|
#0047 add spider
@jpt-c you're gonna be so proud, fancy fancy footwork happening in this one and i was able to do it on my own. done.
Summary
Issue: #0047
Replace "ISSUE_NUMBER" with the number of your issue so that GitHub will link this pull request with the issue and make review easier.
Checklist
All checks are run in GitHub Actions. You'll be able to see the results of the checks at the bottom of the pull request page after it's been opened, and you can click on any of the specific checks listed to see the output of each step and debug failures.
[ ] Tests are implemented
[ ] All tests are passing
[ ] Style checks run (see documentation for more details)
[ ] Style checks are passing
[ ] Code comments from template removed
Questions
Include any questions you have about what you're working on.
๐ very nicely done
|
2025-04-01T04:10:16.743660
| 2022-12-26T04:14:54
|
1510579692
|
{
"authors": [
"codecov-commenter",
"sirtawast"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13790",
"repo": "City-of-Helsinki/yjdh",
"url": "https://github.com/City-of-Helsinki/yjdh/pull/1667"
}
|
gharchive/pull-request
|
[Snyk] Security upgrade next from 11.1.4 to 12.0.9
This PR was automatically created by Snyk using the credentials of a real user.Snyk has created this PR to fix one or more vulnerable packages in the `yarn` dependencies of this project.
Changes included in this PR
Changes to the following files to upgrade the vulnerable dependencies to a fixed version:
frontend/tet/shared/package.json
โ ๏ธ Warning
Failed to update the yarn.lock, please update manually before merging.
Vulnerabilities that will be fixed
With an upgrade:
Severity
Priority Score (*)
Issue
Breaking Change
Exploit Maturity
713/1000 Why? Proof of Concept exploit, Recently disclosed, Has a fix available, CVSS 6.4
Prototype Pollution SNYK-JS-JSON5-3182856
Yes
Proof of Concept
(*) Note that the real score may have changed since the PR was raised.
Check the changes in this PR to ensure they won't cause issues with your project.
Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs.
For more information:
๐ง View latest project report
๐ Adjust project settings
๐ Read more about Snyk's upgrade and patch logic
Learn how to fix vulnerabilities with free interactive lessons:
๐ฆ Prototype Pollution
Codecov Report
Merging #1667 (289dbac) into develop (a4f1222) will decrease coverage by 30.75%.
The diff coverage is n/a.
@@ Coverage Diff @@
## develop #1667 +/- ##
============================================
- Coverage 65.92% 35.17% -30.76%
============================================
Files 1050 82 -968
Lines 29010 1214 -27796
Branches 2972 294 -2678
============================================
- Hits 19126 427 -18699
+ Misses 9436 764 -8672
+ Partials 448 23 -425
Impacted Files
Coverage ฮ
...end/kesaseteli/companies/tests/test_company_api.py
...ared/oidc/migrations/0002_eauthorizationprofile.py
...forms/application/step4/useApplicationFormStep4.ts
frontend/shared/src/utils/object.utils.ts
...plicant/browser-tests/page-model/TermsOfService.ts
backend/shared/shared/oidc/auth.py
...ackend/shared/shared/oidc/tests/test_oidc_views.py
frontend/shared/src/components/table/utils.tsx
...r/src/hooks/application/useApplicationFormField.ts
frontend/kesaseteli/youth/.eslintrc.js
... and 958 more
:mega: Weโre building smart automated test selection to slash your CI/CD build times. Learn more
|
2025-04-01T04:10:16.770096
| 2022-04-06T20:24:33
|
1195145390
|
{
"authors": [
"adriencyberspace",
"kathyccheng"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13791",
"repo": "CityOfDetroit/bloom",
"url": "https://github.com/CityOfDetroit/bloom/issues/1121"
}
|
gharchive/issue
|
Detroit - Fix Arabic layout issue
Acceptance Criteria
Fix the layout issue in Arabic on a listing detail page - right sidebar column should be on the left
Fix the page header component that does not switch orientation
Note: Make sure 'ar' is listed in your public .env LANGUAGES
Entry point 1: http://localhost:3000/about
Choose Arabic as your language
Notice that the page header is on the right instead of left side of the page
Entry point 2: http://localhost:3000/listings
Choose Arabic
Click any listing
Notice that the info container is on the right side of the listing image as opposed to the left
NOTE: This issue only addresses the functional layout of the page. I think it would be worthwhile to create another issue(s) that addresses some funky margin issues by swapping some of the tailwind css tags with the tailwind rtl css tags, i.e. so that an element that normally has margin-right, would then have margin-left when the layout switched to right-to-left.
@adriencyberspace I don't know if this is from your issue or something else, but on dev when I click on any of the other languages on a listing detail page, I get what I think is a page not found error
@kathyccheng I checked with Emily and she said it's a pre-existing issue with a fix that hasn't been deployed yet, so maybe hold off on checking this for a day or two.
LGTM
|
2025-04-01T04:10:16.798964
| 2016-03-01T21:45:05
|
137695976
|
{
"authors": [
"Maxopoly",
"rourke750"
],
"license": "bsd-3-clause",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13792",
"repo": "Civcraft/PrisonPearl",
"url": "https://github.com/Civcraft/PrisonPearl/issues/97"
}
|
gharchive/issue
|
Stack on locate any
Caused by: java.lang.NullPointerException โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท at vg.civcraft.mc.prisonpearl.managers.PrisonPearlManager.freePearlFromMercury(PrisonPearlManager.java:407) ~[?:?] โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท at vg.civcraft.mc.prisonpearl.listener.MercuryListener.pearlUpdate(MercuryListener.java:110) ~[?:?] โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท
at vg.civcraft.mc.prisonpearl.listener.MercuryListener.mercuryListener(MercuryListener.java:53) ~[?:?] โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท at sun.reflect.GeneratedMethodAccessor47.invoke(Unknown Source) ~[?:?] โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_72-internal] โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_72-internal] โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท at org.bukkit.plugin.java.JavaPluginLoader$1.execute(JavaPluginLoader.java:306) ~[spigot.jar:git-Spigot-db6de12-880a532] โยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยทยท ... 11 more
https://github.com/Civcraft/PrisonPearl/commit/38376b0929374c6266a022ffed7075e88a49b062
|
2025-04-01T04:10:16.813937
| 2022-12-09T16:14:33
|
1487017807
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13793",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/1518"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.nl has degraded performance
In f15e8f8, Bofrost.nl (https://www.bofrost.nl) experienced degraded performance:
HTTP code: 200
Response time: 1232 ms
Resolved: Bofrost.nl performance has improved in 8b472b0.
|
2025-04-01T04:10:16.817167
| 2023-07-20T02:46:53
|
1813044660
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13794",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/17659"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.fr has degraded performance
In f5d51b3, Bofrost.fr (https://www.bofrost.fr) experienced degraded performance:
HTTP code: 200
Response time: 1706 ms
Resolved: Bofrost.fr performance has improved in 15fd2ef.
|
2025-04-01T04:10:16.820152
| 2023-08-22T04:46:15
|
1860538935
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13795",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/21099"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.es has degraded performance
In ae763c9, Bofrost.es (https://www.bofrost.es) experienced degraded performance:
HTTP code: 200
Response time: 1516 ms
Resolved: Bofrost.es performance has improved in cc313a4 after 306 days, 16 hours, 20 minutes.
|
2025-04-01T04:10:16.823298
| 2023-10-28T11:38:15
|
1966561599
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13796",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/26250"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.at has degraded performance
In 202c59e, Bofrost.at (https://www.bofrost.at) experienced degraded performance:
HTTP code: 200
Response time: 2476 ms
Resolved: Bofrost.at performance has improved in 0aefb9f after 6 minutes.
|
2025-04-01T04:10:16.826249
| 2023-11-23T16:51:52
|
2008567887
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13797",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/29354"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.lu has degraded performance
In b62debc, Bofrost.lu (https://www.bofrost.lu) experienced degraded performance:
HTTP code: 200
Response time: 3229 ms
Resolved: Bofrost.lu performance has improved in 70c1e0d after 6 minutes.
|
2025-04-01T04:10:16.829692
| 2023-12-05T02:42:07
|
2025163297
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13798",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/30854"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.at has degraded performance
In 8216816, Bofrost.at (https://www.bofrost.at) experienced degraded performance:
HTTP code: 200
Response time: 988 ms
Resolved: Bofrost.at performance has improved in 81ef0b1 after 6 minutes.
|
2025-04-01T04:10:16.832580
| 2023-12-12T10:04:38
|
2037405449
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13799",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/31662"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.es has degraded performance
In c51486e, Bofrost.es (https://www.bofrost.es) experienced degraded performance:
HTTP code: 200
Response time: 1649 ms
Resolved: Bofrost.es performance has improved in ff8a6f8 after 9 minutes.
|
2025-04-01T04:10:16.835498
| 2023-12-18T18:43:15
|
2047230822
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13800",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/32336"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.lu has degraded performance
In cd4968e, Bofrost.lu (https://www.bofrost.lu) experienced degraded performance:
HTTP code: 200
Response time: 1883 ms
Resolved: Bofrost.lu performance has improved in 803928e after 6 minutes.
|
2025-04-01T04:10:16.838429
| 2024-01-04T11:00:31
|
2065442117
|
{
"authors": [
"Ckal"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:13801",
"repo": "Ckal/Uptime",
"url": "https://github.com/Ckal/Uptime/issues/33430"
}
|
gharchive/issue
|
โ ๏ธ Bofrost.lu has degraded performance
In d20927a, Bofrost.lu (https://www.bofrost.lu) experienced degraded performance:
HTTP code: 200
Response time: 1802 ms
Resolved: Bofrost.lu performance has improved in 9fe44a6 after 8 minutes.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.