id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
268132415 | Release 3.0
Branch for last pre-release fixes. As this is major release we can look for some stuff to change. I don't see anything else other than what I updated in these few commits.
Do you have anything to add here @olleolleolle ? I added you as author to gemspec as you're the maintainer. Feel free to revert if you don't like it.
@sickill Thanks for adding my name to the credits list. I'm OK with it. I'm a maintainer, so we're 2 now.
Yes, I often mistake the for a ;)
@sickill Now, if we're targetting 2.1+, we have support for Refinements. Then, the Refinements PR could be brushed up and added.
Aha, that's good idea!
Also realized that having refinements we can drop the legacy String mixin support completely.
@sickill Would it be right to drop in the Refinements thing now (during a major update), and force people to change their usage, or is that a lot to ask of people which depend on this library? (The cost of being popular: can't change. "Avoid success at all costs" --Simon Peyton Jones)
Hmm, good point about popularity. It's not harmful to keep the old String mixin. What we could do:
add refinements in 3.0
keep the mixin in 3.x line, so it works for existing codebases, but remove it from README now so new people only use refinements
remove mixin in 4.0
Sounds good?
@sickill That sounds prudent, 👍
| gharchive/pull-request | 2017-10-24T18:09:22 | 2025-04-01T06:45:47.317205 | {
"authors": [
"olleolleolle",
"sickill"
],
"repo": "sickill/rainbow",
"url": "https://github.com/sickill/rainbow/pull/72",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
409778418 | Codesponsor link at bottom of report broken
Do you want to request a feature or report a bug?
Report a bug
What is the current behavior?
If the current behavior is a bug, please provide the steps to reproduce.
Go to a report page and scroll to the bottom
What is the expected behavior?
Should show an image and link to something? I think codesponsor is now 'codefund' https://codefund.io/ - Oh haha and I just read the end of this template and it mentions opencollective. So maybe it should be updated to that?
Removed the text in https://github.com/siddharthkp/bundlesize/commit/7bced6e6d8279e632fe7b3826bc20263093e1cbd#diff-c4b5d4a975caa9594c8f2232afd745dd, forgot to close this issue
| gharchive/issue | 2019-02-13T12:19:57 | 2025-04-01T06:45:47.322661 | {
"authors": [
"siddharthkp",
"zaccolley"
],
"repo": "siddharthkp/bundlesize",
"url": "https://github.com/siddharthkp/bundlesize/issues/277",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
803369395 | [PROPOSAL] Adding calico typha for eks
Hi team, we are using in a lot of clusters this kustomization module to add Network Policy capabilities in EKS clusters.
Should we add it on this repository?
🤔 Does't calico implement networkpolicies without typha?
| gharchive/pull-request | 2021-02-08T09:23:09 | 2025-04-01T06:45:47.393224 | {
"authors": [
"angelbarrera92",
"nutellinoit"
],
"repo": "sighupio/fury-kubernetes-networking",
"url": "https://github.com/sighupio/fury-kubernetes-networking/pull/20",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
864118086 | include kafkametricsreceiver to collect and ingest kafka consumer lag
This receiver was added by Splunk to be able to collect and ingest consumer lag from kafka. Please let me know if you require changes to go.mod or should I only include changes to the components file.
You'll want to wait for #314 to be merged then rebase on that. https://github.com/signalfx/splunk-otel-collector/blob/main/docs/components.md will need updating as well.
ack, thanks Jay! Will do. we can close this one.
| gharchive/pull-request | 2021-04-21T17:35:58 | 2025-04-01T06:45:47.470063 | {
"authors": [
"harnitsignalfx",
"jrcamp"
],
"repo": "signalfx/splunk-otel-collector",
"url": "https://github.com/signalfx/splunk-otel-collector/pull/315",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2443938223 | Fix config_server_test flakiness
The test flakiness is due to the release of the port used by the ConfigServer not being deterministic since it happens in a goroutine, not during the calls to OnShutdown.
Made the test reliable by actually checking if the port is available.
I'm opting to change the test to match the actual behavior of the type, but, in general we should make resource release deterministic. At least now this non-determinism is being documented.
| gharchive/pull-request | 2024-08-02T03:41:11 | 2025-04-01T06:45:47.471400 | {
"authors": [
"pjanotti"
],
"repo": "signalfx/splunk-otel-collector",
"url": "https://github.com/signalfx/splunk-otel-collector/pull/5175",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
789423146 | Is there pre-processing of data required?
Orion version: 1.0.5
Python version: 3.7.3
Operating System: Windows 10
Description
Describe what you were trying to get done.
Tell us what happened, what went wrong, and what you expected to happen.
Using the example but passing different data:
train_data.head()
timestamp value
0 19941205 3490.1
1 19941212 3488.7
2 19941219 3485.8
3 19941226 3486.1
4 19950102 3480.3
What I Did
orion.fit(train_data)
Paste the command(s) you ran and the output.
If there was a crash, please include the traceback here.
E0119 14:11:53.885504 28744 mlpipeline.py:552] Exception caught fitting MLBlock keras.Sequential.LSTMTimeSeriesRegressor#1
Traceback (most recent call last):
File "c:\usr\Python37\lib\site-packages\mlblocks\mlpipeline.py", line 549, in _fit_block
block.fit(**fit_args)
File "c:\usr\Python37\lib\site-packages\mlblocks\mlblock.py", line 302, in fit
getattr(self.instance, self.fit_method)(**fit_kwargs)
File "c:\usr\Python37\lib\site-packages\mlprimitives\adapters\keras.py", line 98, in fit
self._augment_hyperparameters(X, kwargs)
File "c:\usr\Python37\lib\site-packages\mlprimitives\adapters\keras.py", line 89, in _augment_hyperparameters
shape = np.asarray(X)[0].shape
IndexError: index 0 is out of bounds for axis 0 with size 0
Traceback (most recent call last):
File "", line 1, in
File "c:\usr\Python37\lib\site-packages\orion\core.py", line 84, in fit
self._mlpipeline.fit(data)
File "c:\usr\Python37\lib\site-packages\mlblocks\mlpipeline.py", line 649, in fit
self._fit_block(block, block_name, context)
File "c:\usr\Python37\lib\site-packages\mlblocks\mlpipeline.py", line 549, in _fit_block
block.fit(**fit_args)
File "c:\usr\Python37\lib\site-packages\mlblocks\mlblock.py", line 302, in fit
getattr(self.instance, self.fit_method)(**fit_kwargs)
File "c:\usr\Python37\lib\site-packages\mlprimitives\adapters\keras.py", line 98, in fit
self._augment_hyperparameters(X, kwargs)
File "c:\usr\Python37\lib\site-packages\mlprimitives\adapters\keras.py", line 89, in _augment_hyperparameters
shape = np.asarray(X)[0].shape
IndexError: index 0 is out of bounds for axis 0 with size 0
Hi @juliabelford!
The train_data you are showing does not require additional preprocessing if you are using one of our pipelines. Since you are using your own data, just make sure it follows the same format as we do, which is a pd.DataFrame with two columns (timestamp, value), similar to the one shown here
timestamp value
0 1222819200 -0.366359
1 1222840800 -0.394108
2 1222862400 0.403625
3 1222884000 -0.362759
4 1222905600 -0.370746
Additionally, you need to set the correct hyperparameters for your data! Based on the frequency of your data you want to change interval to reflect that, in the default setting interval=21600 which is equivalent to 6 hours. If your data is more frequent (5 minutes) then interval=300 would be most appropriate. The units here are 1 second.
Note that you can actually shrink your signal by aggregating it over too long of a period, which is what I presume happened in this case.
Stitching it all together would show something like (I am assuming you are running lstm_dynamic_threshold pipeline):
hyperparameters = {
"mlprimitives.custom.timeseries_preprocessing.time_segments_average#1": {
"interval": 300
},
'keras.Sequential.LSTMTimeSeriesRegressor#1': {
'epochs': 5,
'verbose': True
}
}
orion = Orion(
pipeline='lstm_dynamic_threshold',
hyperparameters=hyperparameters
)
orion.fit(train_data)
Changing the aggregating interval should fix your problem, but if it hasn't, would you mind giving a bit more detail on the length of the signal you are using and possibly sharing the entire code snippet so that I can try to reproduce it from my side?
Hi @juliabelford!
The train_data you are showing does not require additional preprocessing if you are using one of our pipelines. Since you are using your own data, just make sure it follows the same format as we do, which is a pd.DataFrame with two columns (timestamp, value), similar to the one shown here
timestamp value
0 1222819200 -0.366359
1 1222840800 -0.394108
2 1222862400 0.403625
3 1222884000 -0.362759
4 1222905600 -0.370746
Additionally, you need to set the correct hyperparameters for your data! Based on the frequency of your data you want to change interval to reflect that, in the default setting interval=21600 which is equivalent to 6 hours. If your data is more frequent (5 minutes) then interval=300 would be most appropriate. The units here are 1 second.
Note that you can actually shrink your signal by aggregating it over too long of a period, which is what I presume happened in this case.
Stitching it all together would show something like (I am assuming you are running lstm_dynamic_threshold pipeline):
hyperparameters = {
"mlprimitives.custom.timeseries_preprocessing.time_segments_average#1": {
"interval": 300
},
'keras.Sequential.LSTMTimeSeriesRegressor#1': {
'epochs': 5,
'verbose': True
}
}
orion = Orion(
pipeline='lstm_dynamic_threshold',
hyperparameters=hyperparameters
)
orion.fit(train_data)
Changing the aggregating interval should fix your problem, but if it hasn't, would you mind giving a bit more detail on the length of the signal you are using and possibly sharing the entire code snippet so that I can try to reproduce it from my side?
My data are daily, so does the interval need to be 246060?
My data are daily, so does the interval need to be 246060?
Based on the input you are showing:
timestamp value
0 19941205 3490.1
1 19941212 3488.7
2 19941219 3485.8
3 19941226 3486.1
4 19950102 3480.3
The difference is:
timestamp[1] - timestamp[0] = 7 seconds
timestamp[2] - timestamp[1] = 7 seconds
timestamp[3] - timestamp[2] = 7 seconds
timestamp[4] - timestamp[3] = 8876 seconds
Since the majority of records seem to be 7 seconds apart, I would set interval=7.
Based on the input you are showing:
timestamp value
0 19941205 3490.1
1 19941212 3488.7
2 19941219 3485.8
3 19941226 3486.1
4 19950102 3480.3
The difference is:
timestamp[1] - timestamp[0] = 7 seconds
timestamp[2] - timestamp[1] = 7 seconds
timestamp[3] - timestamp[2] = 7 seconds
timestamp[4] - timestamp[3] = 8876 seconds
Since the majority of records seem to be 7 seconds apart, I would set interval=7.
setting the interval to 7 did not break the code. However, my timestamps are dates and not seconds. So how would I interpret this result? The test data are from 2020-01-01 till 2021-01-13
anomalies
start end severity
0 20209748 20210112 2.870566
setting the interval to 7 did not break the code. However, my timestamps are dates and not seconds. So how would I interpret this result? The test data are from 2020-01-01 till 2021-01-13
anomalies
start end severity
0 20209748 20210112 2.870566
It depends on how did you convert your original timestamps, you should just invert it.
It depends on how did you convert your original timestamps, you should just invert it.
@sarahmish Can you clarify how "interval" is used. Suppose your time series has a consistent sample rate of 6 samples per minute, or every 10 seconds. What does it mean to set the interval to something larger than 10, say 5 minutes like you originally suggested (interval=300)?
@sarahmish Can you clarify how "interval" is used. Suppose your time series has a consistent sample rate of 6 samples per minute, or every 10 seconds. What does it mean to set the interval to something larger than 10, say 5 minutes like you originally suggested (interval=300)?
Hi @kb1ooo, sure!
The purpose of the mlprimitives.custom.timeseries_preprocessing.time_segments_average#1 primitive is to make your signal equi-spaced. If your signal's frequency is every 10 seconds, then the output of the primitive is a time series with a value at every 10 seconds.
Let's say you have the following signal
timestamp value
0 1404165600 10844.0
1 1404165900 8127.0
2 1404166200 6210.0
3 1404166800 3820.0
It's clear that the interval between each timestamp and the other is 300. Though notice there is a gap between instance 2 and 3, there should be a timestamp 1404166500. The primitive then creates a placeholder (nan) and assumes the imputation primitive will take care of it. The output would be [[10844.], [ 8127.], [ 6210.], [ nan], [ 3820.]].
What happens when you pick an interval larger than the frequency of your signal?
In this case, it will take the average of values. Consider the same signal mentioned above and we set the interval to be 600, then we take the average of the interval. In this case the output would be [[9485.5], [6210. ], [3820. ]].
Let me know if you need further clarification
Hi @kb1ooo, sure!
The purpose of the mlprimitives.custom.timeseries_preprocessing.time_segments_average#1 primitive is to make your signal equi-spaced. If your signal's frequency is every 10 seconds, then the output of the primitive is a time series with a value at every 10 seconds.
Let's say you have the following signal
timestamp value
0 1404165600 10844.0
1 1404165900 8127.0
2 1404166200 6210.0
3 1404166800 3820.0
It's clear that the interval between each timestamp and the other is 300. Though notice there is a gap between instance 2 and 3, there should be a timestamp 1404166500. The primitive then creates a placeholder (nan) and assumes the imputation primitive will take care of it. The output would be [[10844.], [ 8127.], [ 6210.], [ nan], [ 3820.]].
What happens when you pick an interval larger than the frequency of your signal?
In this case, it will take the average of values. Consider the same signal mentioned above and we set the interval to be 600, then we take the average of the interval. In this case the output would be [[9485.5], [6210. ], [3820. ]].
Let me know if you need further clarification
@sarahmish Thanks very much for your detailed reply. One further question, what if you have multiple disjoint signals for training, that you don't want to interpolate over? E.g. say you have several 5 day contiguous chunks of data.
Thanks,
Marc
@sarahmish Thanks very much for your detailed reply. One further question, what if you have multiple disjoint signals for training, that you don't want to interpolate over? E.g. say you have several 5 day contiguous chunks of data.
Thanks,
Marc
Interesting case @kb1ooo
Our preprocessing primitives are based on the assumption that the signal is contiguous. So what happens when the data is non-contiguous?
Running the pipeline without any changes
As you would have guessed, the mlprimitives.custom.timeseries_preprocessing.time_segments_average primitive would introduce many nan into your signal, which will then get imputed but the resulting signal will not resemble your original signal and hence, the anomaly detection model might find anomalies because of the imputation artifact.
Removing a primitive
Assume that your signal is already equi-spaced within each chunk. Then technically you do not require the above primitive. However, the mlprimitives.custom.timeseries_preprocessing.rolling_window_sequences primitive will assume continuity and create training windows that interleave between days.
Preprocess each chunk individually
I would recommend applying the preprocessing steps on each disjoint segment individually. Piece the training examples together, and then train the model. Unfortunately, we don't support this right now.
Run the pipeline on each chunk individually
The last option would would to treat each chunk as a separate signal and run the pipeline on each one. In this case, you would be training several models which might not be ideal.
Interesting case @kb1ooo
Our preprocessing primitives are based on the assumption that the signal is contiguous. So what happens when the data is non-contiguous?
Running the pipeline without any changes
As you would have guessed, the mlprimitives.custom.timeseries_preprocessing.time_segments_average primitive would introduce many nan into your signal, which will then get imputed but the resulting signal will not resemble your original signal and hence, the anomaly detection model might find anomalies because of the imputation artifact.
Removing a primitive
Assume that your signal is already equi-spaced within each chunk. Then technically you do not require the above primitive. However, the mlprimitives.custom.timeseries_preprocessing.rolling_window_sequences primitive will assume continuity and create training windows that interleave between days.
Preprocess each chunk individually
I would recommend applying the preprocessing steps on each disjoint segment individually. Piece the training examples together, and then train the model. Unfortunately, we don't support this right now.
Run the pipeline on each chunk individually
The last option would would to treat each chunk as a separate signal and run the pipeline on each one. In this case, you would be training several models which might not be ideal.
@sarahmish Thanks again for the detailed reply. I think your "Preprocess each chunk individually" makes the most sense, and is what I have done in the past for other implementations. For realistic problems, I believe it's quite common to have training data with significant gaps that cannot be imputed. You may even want to deliberately set it up that way for performance evaluation purposes (e.g. suppose you have 12 months of data, you may want to hold out 1 week of random data each month for testing, as opposed to say 10 months of contiguous data for training and 2 months for testing). I'm not sure where the windowing/imputation happens, but it seems that having a "maxgap" parameter which would specify the maximum gap in seconds over which you allow imputation, would make sense. Your signal gets split on >maxgap gaps and fed to the windowing. Can you point me to where I might look to see what those changes would entail?
Thanks,
Marc
@sarahmish Thanks again for the detailed reply. I think your "Preprocess each chunk individually" makes the most sense, and is what I have done in the past for other implementations. For realistic problems, I believe it's quite common to have training data with significant gaps that cannot be imputed. You may even want to deliberately set it up that way for performance evaluation purposes (e.g. suppose you have 12 months of data, you may want to hold out 1 week of random data each month for testing, as opposed to say 10 months of contiguous data for training and 2 months for testing). I'm not sure where the windowing/imputation happens, but it seems that having a "maxgap" parameter which would specify the maximum gap in seconds over which you allow imputation, would make sense. Your signal gets split on >maxgap gaps and fed to the windowing. Can you point me to where I might look to see what those changes would entail?
Thanks,
Marc
@kb1ooo thanks for the suggestion. I opened a new issue relevant to the idea you are suggesting #171.
I listed the primary primitives you would need to investigate there. The approach I am proposing does not alter the current implementation but rather is an additive. However, I believe you can investigate these primitives if you have a different approach; for imputation we directly use sklearn.impute.SimpleImputer and for windowing we use mlprimitives.custom.timeseries_preprocessing.rolling_window_sequence.
@sarahmish thank you, I will take a look per your suggestions in #171
| gharchive/issue | 2021-01-19T22:20:40 | 2025-04-01T06:45:47.509083 | {
"authors": [
"juliabelford",
"kb1ooo",
"sarahmish"
],
"repo": "signals-dev/Orion",
"url": "https://github.com/signals-dev/Orion/issues/162",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2525940082 | 🛑 sfgz is down
In 1373da0, sfgz (https://sfgz.ch/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: sfgz is back up in a387a61 after 1 hour, 53 minutes.
| gharchive/issue | 2024-09-14T02:27:17 | 2025-04-01T06:45:47.512422 | {
"authors": [
"signalwerk"
],
"repo": "signalwerk/uptime",
"url": "https://github.com/signalwerk/uptime/issues/817",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1661992357 | Add support for Azure OpenAI
This PR adds support for the Azure OpenAI service, which is an alternate API for accessing the OpenAI models. In the initial-config wizard it adds a question Use OpenAI via Azure? (default no), and if the answer is yes it asks for the endpoint and deployment name. It stores these in the config file under new keys. It also adds support for max_tokens (on both APIs), since the Azure max_tokens default is very low.
Details:
If aoai_endpoint is set, then use Azure OpenAI endpoint rather than the direct OpenAI endpoint. Requires aoai_deployment to be specified as well. You should set the model too, so that aichat knows the correct number of remaining tokens.
Also adds support for max_tokens since the AOAI default is very low.
The Azure OpenAI API currently requires chats to be encoded as completions, so logic to do this is added as well.
Tested with gpt-35-turbo (version 0301).
After pr #150, aichat supports multiple clients.
It would be better to add an independent azure-openai client. @kw217 @FalacerSelene
see #166
| gharchive/pull-request | 2023-04-11T08:18:35 | 2025-04-01T06:45:47.537177 | {
"authors": [
"kw217",
"sigoden"
],
"repo": "sigoden/aichat",
"url": "https://github.com/sigoden/aichat/pull/106",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1797014104 | Figure out a better way to keep the beacon alive
The beacon currently runs on a schedule; GitHub Actions disables scheduled workflows after 60 days of repository inactivity.
As a result: if nobody commits to this repo in a 60 day window, the beacon's scheduled job stops running and repos that test using the Sigstore conformance suite begin to fail. For example, the timeout failures in https://github.com/sigstore/sigstore-python/pull/702.
Some ideas:
Maybe there's some way to get this repository excluded from the scheduled workflow expiration rule. CC @steiza?
Maybe we can "cheese" things here and have a periodic job that also commits to the repo, to keep the workflows active.
As a resolution here, we'll start with (2).
Happened again: https://github.com/sigstore/sigstore-conformance/issues/117
Maybe we can "cheese" things here and have a periodic job that also commits to the repo
My previous attempt at making the token more easily available via Pages failed (see #8) but I had another idea for that: commiting the current token to git and maybe force pushing a tag/branch that points to the current-token-commit (re-using the tag/branch to avoid polluting git checkouts too much). It sounds like a hack but it might be a two-birds-with-a-single-commit type of hack...
I like that idea!
(@steiza: do you happen to know if machine commits to a branch count for resetting the workflow expiration countdown? I know Dependabot commits/PRs do, but that might only be because they get merged by a human.)
I ran into this when monitoring the log, because the repository only contains a workflow that never commits. I found https://github.com/marketplace/actions/keepalive-workflow as a potential solution, though I haven't tried it out yet.
I don't know yet if #17 makes sense but it might accidentally fix this one
| gharchive/issue | 2023-07-10T15:24:58 | 2025-04-01T06:45:47.543798 | {
"authors": [
"haydentherapper",
"jku",
"woodruffw"
],
"repo": "sigstore-conformance/extremely-dangerous-public-oidc-beacon",
"url": "https://github.com/sigstore-conformance/extremely-dangerous-public-oidc-beacon/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1175907208 | Binary for Debian release has unexpected name
Description
Not entirely sure if this is intentional or not, but I was surprised to find that installing the Debian release (added in #1537) creates a binary named cosign-linux-amd64 and not cosign.
$ docker run -it debian
root@3fe738d1ed8b:/# apt-get update && apt-get install wget
...
root@3fe738d1ed8b:/# wget -q https://github.com/sigstore/cosign/releases/download/v1.6.0/cosign_1.6.0_amd64.deb
root@3fe738d1ed8b:/# dpkg -i cosign_1.6.0_amd64.deb
Selecting previously unselected package cosign.
(Reading database ... 7071 files and directories currently installed.)
Preparing to unpack cosign_1.6.0_amd64.deb ...
Unpacking cosign (1.6.0) ...
Setting up cosign (1.6.0) ...
root@3fe738d1ed8b:/# dpkg -L cosign
/usr
/usr/local
/usr/local/bin
/usr/local/bin/cosign-linux-amd64
/usr/local/bin/cosign-linux-pivkey-pkcs11key-amd64
Figured I'd report this in case it was unintentional. Haven't tested any of the other releases from #1537 but this behavior might be present there as well.
(cc @strongjz)
This is probably a quirk in the goreleaser setup. Agree it's a bug.
I will kook into the templating for the name.
Looks like .ConventionalFileName | conventional package file name as provided by nFPM does just that, copies the name from the build. which is cosign-linux-{{ .Arch }} We can probably just use trimsuffix to remove anything after the -.
The note/issue for arm version doesn't seem to apply since we specify goarm 7
https://goreleaser.com/customization/templates/?h=conventionalfilename#fn:9
Please beware: some OSs might have the same names for different ARM versions, for example, for Debian both ARMv6 and ARMv7 are called armhf. Make sure that's not your case otherwise you might end up with colliding names.
https://github.com/golang/go/wiki/GoArm
https://github.com/goreleaser/goreleaser/issues/36
This is still appears to be the case with v1.7.1 deb:
Selecting previously unselected package cosign.
(Reading database ... 212247 files and directories currently installed.)
Preparing to unpack cosign.deb ...
Unpacking cosign (1.7.1) ...
Setting up cosign (1.7.1) ...
~
❯ dpkg -L cosign
/usr
/usr/bin
/usr/bin/cosign-linux-amd64
/usr/bin/cosign-linux-pivkey-pkcs11key-amd64
looks like we changed only the package name but not in binary name inside the package. I think this is not possible to change.
also changing the name broke the release because we now have the same package name for all architectures ( https://github.com/sigstore/cosign/issues/1733 )
Hi, I have encounter this issue.
Maybe it will be nice to add the extra steps in the documentation for the new comers.
I renamed the binary folder but I am not sure if it is the right thing to do. Will it be adding complexity for the future upgrade?
| gharchive/issue | 2022-03-21T20:29:32 | 2025-04-01T06:45:47.550536 | {
"authors": [
"Dwynie",
"cpanato",
"di",
"dlorenc",
"strongjz",
"tekumara"
],
"repo": "sigstore/cosign",
"url": "https://github.com/sigstore/cosign/issues/1643",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
975033745 | Fix output display for certs.
Signed-off-by: Dan Lorenc dlorenc@google.com
@dlorenc do you prefer to tackle it here? Cos this is changing the same thing I was working on https://github.com/sigstore/cosign/pull/546, sorry for the delay answering there, the ebpf-summit was this week I was hook on it.
Let's get yours in! Happy to close mine. I forgot about this fix!
| gharchive/pull-request | 2021-08-19T20:43:52 | 2025-04-01T06:45:47.552820 | {
"authors": [
"dlorenc",
"josedonizetti"
],
"repo": "sigstore/cosign",
"url": "https://github.com/sigstore/cosign/pull/553",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1348503222 | v1.fulcio.sigstore.dev returns k8s ingress controller fake cert
Description
v1.fulcio.sigstore.dev fails TLS validation because it returns a k8s cert:
Subject: O = Acme Co, CN = Kubernetes Ingress Controller Fake Certificate
fulcio.sigstore.dev does return a valid cert with alt names for fulcio.sigstore.dev and v1.fulcio.sigstore.dev
v1.fulcio.sigstore.dev:
$ openssl s_client -connect v1.fulcio.sigstore.dev:443 | openssl x509 -text
depth=0 O = Acme Co, CN = Kubernetes Ingress Controller Fake Certificate
verify error:num=18:self-signed certificate
verify return:1
depth=0 O = Acme Co, CN = Kubernetes Ingress Controller Fake Certificate
verify return:1
Certificate:
Data:
Version: 3 (0x2)
Serial Number:
c5:ba:71:95:75:df:73:10:53:ad:86:67:6d:aa:9c:20
Signature Algorithm: sha256WithRSAEncryption
Issuer: O = Acme Co, CN = Kubernetes Ingress Controller Fake Certificate
Validity
Not Before: Aug 17 07:24:15 2022 GMT
Not After : Aug 17 07:24:15 2023 GMT
Subject: O = Acme Co, CN = Kubernetes Ingress Controller Fake Certificate
Subject Public Key Info:
Public Key Algorithm: rsaEncryption
Public-Key: (2048 bit)
Modulus:
00:c3:75:96:9e:30:03:8a:97:cd:85:04:54:d0:31:
a2:3c:9d:6b:f7:c0:f1:fc:e2:6a:3a:42:7f:24:05:
fb:b3:31:58:79:a5:d8:57:0d:e4:91:d3:72:b2:05:
ec:ac:3d:d0:3d:74:0c:fc:e3:dc:b4:28:c6:f1:83:
36:1f:a7:8a:d1:89:f0:00:67:a5:74:18:ac:a5:fa:
5f:4e:f2:2b:c2:e4:c4:0b:9e:44:9c:89:c2:0a:65:
37:9e:e4:bc:df:e8:b9:c1:c3:0f:21:73:93:69:db:
40:91:0e:b3:61:82:0c:03:15:6b:21:ea:88:17:c1:
e5:07:48:20:1b:58:61:3f:da:a9:69:03:7c:5a:dc:
ae:6e:ba:20:d8:2f:d5:69:21:af:2e:4a:43:08:3c:
2b:7b:34:0c:11:82:60:e8:5e:6d:54:ba:c7:33:fc:
0e:ea:81:0a:1c:ad:42:a3:0e:bb:e2:68:75:93:a1:
e4:5b:bd:27:34:b1:0e:52:72:a6:48:64:fc:d0:49:
40:3b:4c:a9:3c:5a:88:22:c6:9c:61:e0:c2:82:8b:
0f:b5:ae:ba:a2:fb:9f:36:a4:a6:37:c0:5b:ae:04:
d8:86:77:81:73:c3:b5:58:cc:b9:f4:66:d5:81:fa:
63:81:40:a5:e5:d7:4b:0c:d5:70:d9:d8:3c:bb:9f:
b3:97
Exponent: 65537 (0x10001)
X509v3 extensions:
X509v3 Key Usage: critical
Digital Signature, Key Encipherment
X509v3 Extended Key Usage:
TLS Web Server Authentication
X509v3 Basic Constraints: critical
CA:FALSE
X509v3 Subject Alternative Name:
DNS:ingress.local
Signature Algorithm: sha256WithRSAEncryption
Signature Value:
bb:4d:20:74:a6:04:15:9b:fb:89:bf:1c:22:4a:7f:1c:ae:0e:
31:7e:c5:61:4d:52:c9:83:2d:22:34:bc:24:ac:7f:b7:d0:3f:
5f:4e:b1:c1:85:77:6a:86:c6:fe:5d:18:38:28:33:c3:7b:97:
b0:d0:71:2b:26:48:9a:c2:5b:4a:91:fe:ac:3b:cf:d2:e0:06:
3f:a7:ab:14:79:cc:ec:dc:39:5a:53:f0:b9:b3:61:46:b6:86:
fe:06:a1:f7:10:07:a0:60:c0:ba:36:9d:84:10:6d:56:05:36:
13:b7:26:55:05:8b:a0:61:4f:30:57:ec:2e:43:38:35:85:95:
49:4e:71:d8:e1:67:ee:0c:0f:df:08:c0:ea:bb:f2:2a:33:a9:
80:b0:4e:c7:8b:8e:42:31:c0:5a:0e:cf:cd:89:49:58:48:97:
19:5e:46:2b:2a:68:31:bf:bf:dc:84:4d:9c:16:54:ad:5d:5a:
e0:c2:a3:88:06:63:d0:8e:cd:87:ad:1e:76:c5:07:cb:34:d6:
e7:d3:9d:84:3d:4a:e1:0f:80:d2:43:52:c6:6e:53:4a:33:38:
d0:d1:ff:e0:e7:f5:55:8a:9b:ca:9c:8f:0e:c6:bd:86:b9:ae:
ae:4a:33:36:ad:91:b4:43:29:c7:26:cd:a4:17:db:31:26:00:
2a:64:69:b8
-----BEGIN CERTIFICATE-----
MIIDcDCCAligAwIBAgIRAMW6cZV133MQU62GZ22qnCAwDQYJKoZIhvcNAQELBQAw
SzEQMA4GA1UEChMHQWNtZSBDbzE3MDUGA1UEAxMuS3ViZXJuZXRlcyBJbmdyZXNz
IENvbnRyb2xsZXIgRmFrZSBDZXJ0aWZpY2F0ZTAeFw0yMjA4MTcwNzI0MTVaFw0y
MzA4MTcwNzI0MTVaMEsxEDAOBgNVBAoTB0FjbWUgQ28xNzA1BgNVBAMTLkt1YmVy
bmV0ZXMgSW5ncmVzcyBDb250cm9sbGVyIEZha2UgQ2VydGlmaWNhdGUwggEiMA0G
CSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDDdZaeMAOKl82FBFTQMaI8nWv3wPH8
4mo6Qn8kBfuzMVh5pdhXDeSR03KyBeysPdA9dAz849y0KMbxgzYfp4rRifAAZ6V0
GKyl+l9O8ivC5MQLnkScicIKZTee5Lzf6LnBww8hc5Np20CRDrNhggwDFWsh6ogX
weUHSCAbWGE/2qlpA3xa3K5uuiDYL9VpIa8uSkMIPCt7NAwRgmDoXm1Uuscz/A7q
gQocrUKjDrviaHWToeRbvSc0sQ5ScqZIZPzQSUA7TKk8Wogixpxh4MKCiw+1rrqi
+582pKY3wFuuBNiGd4Fzw7VYzLn0ZtWB+mOBQKXl10sM1XDZ2Dy7n7OXAgMBAAGj
TzBNMA4GA1UdDwEB/wQEAwIFoDATBgNVHSUEDDAKBggrBgEFBQcDATAMBgNVHRMB
Af8EAjAAMBgGA1UdEQQRMA+CDWluZ3Jlc3MubG9jYWwwDQYJKoZIhvcNAQELBQAD
ggEBALtNIHSmBBWb+4m/HCJKfxyuDjF+xWFNUsmDLSI0vCSsf7fQP19OscGFd2qG
xv5dGDgoM8N7l7DQcSsmSJrCW0qR/qw7z9LgBj+nqxR5zOzcOVpT8LmzYUa2hv4G
ofcQB6BgwLo2nYQQbVYFNhO3JlUFi6BhTzBX7C5DODWFlUlOcdjhZ+4MD98IwOq7
8iozqYCwTseLjkIxwFoOz82JSVhIlxleRisqaDG/v9yETZwWVK1dWuDCo4gGY9CO
zYetHnbFB8s01ufTnYQ9SuEPgNJDUsZuU0ozONDR/+Dn9VWKm8qcjw7GvYa5rq5K
MzatkbRDKccmzaQX2zEmACpkabg=
-----END CERTIFICATE-----
fulcio.sigstore.dev:
openssl s_client -connect fulcio.sigstore.dev:443 | openssl x509 -text
depth=2 C = US, O = Internet Security Research Group, CN = ISRG Root X1
verify return:1
depth=1 C = US, O = Let's Encrypt, CN = R3
verify return:1
depth=0 CN = fulcio.sigstore.dev
verify return:1
Certificate:
Data:
Version: 3 (0x2)
Serial Number:
03:a0:51:ae:23:d1:06:32:ea:89:db:ea:1b:48:53:93:60:d2
Signature Algorithm: sha256WithRSAEncryption
Issuer: C = US, O = Let's Encrypt, CN = R3
Validity
Not Before: Aug 23 18:45:37 2022 GMT
Not After : Nov 21 18:45:36 2022 GMT
Subject: CN = fulcio.sigstore.dev
Subject Public Key Info:
Public Key Algorithm: rsaEncryption
Public-Key: (2048 bit)
Modulus:
00:d3:ca:80:3c:b0:92:74:fe:c1:5e:75:6e:74:e7:
b7:3c:9b:a5:7e:3a:3f:30:ea:88:70:c2:7c:60:3f:
09:7c:6c:1e:e1:4a:a6:ca:fd:cd:b2:73:d9:e9:41:
d9:70:32:0d:20:c8:e9:e0:71:f8:f9:78:63:1b:b7:
00:da:a2:09:dd:64:4b:de:2f:62:5c:2c:58:60:ed:
ea:13:bb:d0:f6:6d:7c:34:ca:95:31:3e:b6:d2:e0:
a1:72:8f:2e:08:b4:92:97:87:b8:fc:30:5f:a1:40:
f7:32:17:f8:9f:3e:b1:24:dd:9a:ad:4a:31:71:d3:
dd:d1:0b:c7:24:c7:e5:8d:ce:6c:71:ad:1b:2f:e9:
72:d1:5d:99:a0:da:2c:85:37:27:bc:9f:a9:b5:34:
3f:33:27:e6:02:33:5a:b8:a7:b2:3e:6c:d7:8f:11:
2b:3b:18:48:5a:f0:f4:82:11:ea:2f:c8:ae:c4:8d:
45:39:fa:4c:52:96:bc:53:1b:fe:80:ad:08:f8:18:
cd:7c:7a:35:d6:9c:67:12:34:7d:19:ed:fc:94:ec:
6d:d5:37:27:a2:fb:aa:a0:8b:b5:da:eb:e5:86:fc:
8e:17:87:01:2f:bb:91:1f:45:66:17:c8:65:34:dc:
62:57:b3:56:ee:a2:a8:4a:52:8b:fa:9d:29:e5:d6:
24:69
Exponent: 65537 (0x10001)
X509v3 extensions:
X509v3 Key Usage: critical
Digital Signature, Key Encipherment
X509v3 Extended Key Usage:
TLS Web Server Authentication, TLS Web Client Authentication
X509v3 Basic Constraints: critical
CA:FALSE
X509v3 Subject Key Identifier:
F9:1C:A1:F5:8C:8B:0C:D0:81:AB:73:A9:81:5D:54:A9:9A:54:3E:E8
X509v3 Authority Key Identifier:
14:2E:B3:17:B7:58:56:CB:AE:50:09:40:E6:1F:AF:9D:8B:14:C2:C6
Authority Information Access:
OCSP - URI:http://r3.o.lencr.org
CA Issuers - URI:http://r3.i.lencr.org/
X509v3 Subject Alternative Name:
DNS:fulcio.sigstore.dev, DNS:v1.fulcio.sigstore.dev
X509v3 Certificate Policies:
Policy: 2.23.140.1.2.1
Policy: 1.3.6.1.4.1.44947.1.1.1
CPS: http://cps.letsencrypt.org
CT Precertificate SCTs:
Signed Certificate Timestamp:
Version : v1 (0x0)
Log ID : 41:C8:CA:B1:DF:22:46:4A:10:C6:A1:3A:09:42:87:5E:
4E:31:8B:1B:03:EB:EB:4B:C7:68:F0:90:62:96:06:F6
Timestamp : Aug 23 19:45:37.388 2022 GMT
Extensions: none
Signature : ecdsa-with-SHA256
30:45:02:21:00:E0:B1:8E:D7:95:66:56:9F:3C:64:D0:
B9:49:32:24:45:2B:78:94:AD:27:1E:20:9A:5B:EB:1B:
F8:5B:5E:FC:C8:02:20:18:90:89:F2:B7:0A:5C:C6:98:
83:7C:C8:B8:5E:16:7E:56:9D:20:9E:6F:48:55:CD:54:
0E:3E:4F:40:34:BF:40
Signed Certificate Timestamp:
Version : v1 (0x0)
Log ID : 29:79:BE:F0:9E:39:39:21:F0:56:73:9F:63:A5:77:E5:
BE:57:7D:9C:60:0A:F8:F9:4D:5D:26:5C:25:5D:C7:84
Timestamp : Aug 23 19:45:37.357 2022 GMT
Extensions: none
Signature : ecdsa-with-SHA256
30:45:02:20:15:6E:0B:D1:31:A1:EC:D3:51:09:5C:18:
92:C4:2C:72:4A:30:D4:58:F7:77:D4:C7:67:4C:93:AC:
E1:BE:E9:1A:02:21:00:EF:69:EB:FD:F2:5D:62:C6:6F:
EC:56:60:67:EB:C5:01:27:64:24:16:FE:2C:1A:85:AF:
AE:DA:73:2B:D6:2B:5A
Signature Algorithm: sha256WithRSAEncryption
Signature Value:
3a:63:5d:65:5d:6b:c7:df:18:b6:83:1b:3c:7d:aa:94:21:3d:
aa:ad:f5:88:6a:62:7a:6d:e2:b8:74:58:fe:8d:ea:4f:3d:52:
19:05:16:1f:89:96:ff:83:ef:43:4c:33:7a:75:70:9f:6f:cd:
d2:c9:0e:78:ad:bf:7b:9c:0b:1c:84:2f:db:2a:7d:35:8a:61:
bc:75:1c:29:7e:4a:c0:7d:1d:bf:18:50:e2:85:ab:a8:18:98:
75:24:2a:9f:90:c8:24:bf:74:94:c4:8c:9b:00:0c:3e:5f:be:
a5:09:27:7d:5d:9e:0f:23:67:b3:c7:6c:11:e0:dd:00:7b:88:
ae:57:90:16:a5:d7:f4:03:f1:c9:ef:a7:88:cf:28:a8:63:2c:
b1:09:e7:f8:9c:4f:a9:71:ce:82:30:2d:b7:44:b2:86:ec:14:
d5:9b:d5:4c:3b:ff:59:12:45:54:81:e1:6e:0c:72:1c:ee:64:
65:a9:66:7a:e7:c2:d2:c6:44:37:87:3f:cf:7d:79:06:e1:63:
3a:ce:a2:3e:9a:db:bc:e4:20:88:8c:7c:0b:19:a9:21:82:d9:
b7:47:9c:37:ea:eb:b8:ce:9d:34:7e:25:e1:ad:40:ba:18:35:
d7:b9:3c:91:52:88:16:44:9c:47:48:7a:5d:12:97:8c:90:46:
69:64:f6:57
-----BEGIN CERTIFICATE-----
MIIFRDCCBCygAwIBAgISA6BRriPRBjLqidvqG0hTk2DSMA0GCSqGSIb3DQEBCwUA
MDIxCzAJBgNVBAYTAlVTMRYwFAYDVQQKEw1MZXQncyBFbmNyeXB0MQswCQYDVQQD
EwJSMzAeFw0yMjA4MjMxODQ1MzdaFw0yMjExMjExODQ1MzZaMB4xHDAaBgNVBAMT
E2Z1bGNpby5zaWdzdG9yZS5kZXYwggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEK
AoIBAQDTyoA8sJJ0/sFedW5057c8m6V+Oj8w6ohwwnxgPwl8bB7hSqbK/c2yc9np
QdlwMg0gyOngcfj5eGMbtwDaogndZEveL2JcLFhg7eoTu9D2bXw0ypUxPrbS4KFy
jy4ItJKXh7j8MF+hQPcyF/ifPrEk3ZqtSjFx093RC8ckx+WNzmxxrRsv6XLRXZmg
2iyFNye8n6m1ND8zJ+YCM1q4p7I+bNePESs7GEha8PSCEeovyK7EjUU5+kxSlrxT
G/6ArQj4GM18ejXWnGcSNH0Z7fyU7G3VNyei+6qgi7Xa6+WG/I4XhwEvu5EfRWYX
yGU03GJXs1buoqhKUov6nSnl1iRpAgMBAAGjggJmMIICYjAOBgNVHQ8BAf8EBAMC
BaAwHQYDVR0lBBYwFAYIKwYBBQUHAwEGCCsGAQUFBwMCMAwGA1UdEwEB/wQCMAAw
HQYDVR0OBBYEFPkcofWMiwzQgatzqYFdVKmaVD7oMB8GA1UdIwQYMBaAFBQusxe3
WFbLrlAJQOYfr52LFMLGMFUGCCsGAQUFBwEBBEkwRzAhBggrBgEFBQcwAYYVaHR0
cDovL3IzLm8ubGVuY3Iub3JnMCIGCCsGAQUFBzAChhZodHRwOi8vcjMuaS5sZW5j
ci5vcmcvMDYGA1UdEQQvMC2CE2Z1bGNpby5zaWdzdG9yZS5kZXaCFnYxLmZ1bGNp
by5zaWdzdG9yZS5kZXYwTAYDVR0gBEUwQzAIBgZngQwBAgEwNwYLKwYBBAGC3xMB
AQEwKDAmBggrBgEFBQcCARYaaHR0cDovL2Nwcy5sZXRzZW5jcnlwdC5vcmcwggEE
BgorBgEEAdZ5AgQCBIH1BIHyAPAAdgBByMqx3yJGShDGoToJQodeTjGLGwPr60vH
aPCQYpYG9gAAAYLMPSxsAAAEAwBHMEUCIQDgsY7XlWZWnzxk0LlJMiRFK3iUrSce
IJpb6xv4W178yAIgGJCJ8rcKXMaYg3zIuF4WfladIJ5vSFXNVA4+T0A0v0AAdgAp
eb7wnjk5IfBWc59jpXflvld9nGAK+PlNXSZcJV3HhAAAAYLMPSxNAAAEAwBHMEUC
IBVuC9ExoezTUQlcGJLELHJKMNRY93fUx2dMk6zhvukaAiEA72nr/fJdYsZv7FZg
Z+vFASdkJBb+LBqFr67acyvWK1owDQYJKoZIhvcNAQELBQADggEBADpjXWVda8ff
GLaDGzx9qpQhPaqt9YhqYnpt4rh0WP6N6k89UhkFFh+Jlv+D70NMM3p1cJ9vzdLJ
Dnitv3ucCxyEL9sqfTWKYbx1HCl+SsB9Hb8YUOKFq6gYmHUkKp+QyCS/dJTEjJsA
DD5fvqUJJ31dng8jZ7PHbBHg3QB7iK5XkBal1/QD8cnvp4jPKKhjLLEJ5/icT6lx
zoIwLbdEsobsFNWb1Uw7/1kSRVSB4W4MchzuZGWpZnrnwtLGRDeHP899eQbhYzrO
oj6a27zkIIiMfAsZqSGC2bdHnDfq67jOnTR+JeGtQLoYNde5PJFSiBZEnEdIel0S
l4yQRmlk9lc=
-----END CERTIFICATE-----
Version
@bobcallaway
Please switch to “fulcio.Sigstore.dev”
Please switch to fulcio.sigstore.dev. v1.fulcio.sigstore.dev was deprecated awhile ago, though we are actively working on adding in routing to avoid breaking anyone using it.
Please switch to fulcio.sigstore.dev. v1.fulcio.sigstore.dev was deprecated awhile ago, though we are actively working on adding in routing to avoid breaking anyone using it.
yes, I am attempted to update the github action to track main (v1.11.0) which hopefully takes care of calling the right endpoint.
I felt it was worth reporting given that the host serving fulcio.sigstore.dev does have v1.fulcio.sigstore.dev in its alternate names.
Thanks you for reporting this! It should be fixed now.
| gharchive/issue | 2022-08-23T20:22:13 | 2025-04-01T06:45:47.560772 | {
"authors": [
"chantra",
"haydentherapper"
],
"repo": "sigstore/fulcio",
"url": "https://github.com/sigstore/fulcio/issues/753",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1197665395 | Add priyawadhwa@ as a reviewer on sigstore/scaffolding
I recently added terraform modules to the repo & it would be nice to have review permissions/write access while iterating on them.
+1
I made @priyawadhwa an writer, lmk if that's not enough and I'll fix.
Seems like I can approve PRs now but still can't merge them
Ok, you're now an admin :)
| gharchive/issue | 2022-04-08T18:30:07 | 2025-04-01T06:45:47.564229 | {
"authors": [
"cpanato",
"priyawadhwa",
"vaikas"
],
"repo": "sigstore/scaffolding",
"url": "https://github.com/sigstore/scaffolding/issues/100",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1422517586 | update website to denote GA
Signed-off-by: Bob Callaway bcallaway@google.com
Summary
Release Note
Documentation
+10000
| gharchive/pull-request | 2022-10-25T14:00:05 | 2025-04-01T06:45:47.566297 | {
"authors": [
"bobcallaway",
"dlorenc"
],
"repo": "sigstore/sigstore-website",
"url": "https://github.com/sigstore/sigstore-website/pull/232",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1704430912 | Show "node" subfields for unset schema params
This PR addresses the fact that our change to storing all value-specific data underneath the "node" dict makes the subfields not appear for unset parameters in a manifest. I fixed this by adding a 'default', 'default' entry that is always empty and is used to initialize the node dict whenever it gets populated, mimicking the pattern that gets used for schema parameter keypaths themselves.
I was originally planning to have 'global', 'global' be always filled in by default, but I realized this mucks with some of the logic as it loses the distinction between required and optional pernode parameters (required ones should not have a 'global', 'global' entry).
Github is currently reporting service degradation for Actions, will re-run once it's restored: https://www.githubstatus.com/
@gadfort I ended up implementing what we discussed offline where the defvalue field gets replaced by the value field under default, default. This change actually felt pretty natural during implementation - it simplifies the code in a few places.
Although this is not what we discussed, I decided not to provide a mechanism to override the default value (but I did add a mechanism to retrieve it). Even though this technically regresses on functionality, I still think it feels a little iffy to do it in the first place, and it might be okay to not allow it unless we come up with a compelling reason to make it possible.
Please update docs and resolve merge conflicts
| gharchive/pull-request | 2023-05-10T18:24:09 | 2025-04-01T06:45:47.580080 | {
"authors": [
"gadfort",
"nmoroze"
],
"repo": "siliconcompiler/siliconcompiler",
"url": "https://github.com/siliconcompiler/siliconcompiler/pull/1544",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1819725740 | Can't create project in k8s locally
When I try to create a new project in k8s locally, it doesn't seem to work and the result is always a 404 on the project page I'm redirected to:
fail: LexBoxApi.GraphQL.ErrorLoggingDiagnosticsEventListener[0]
ResolverError: Unexpected Execution Error
System.Net.Http.HttpRequestException: Response status code does not indicate success: 404 (Not Found).
at System.Net.Http.HttpResponseMessage.EnsureSuccessStatusCode()
at LexBoxApi.Services.HgService.GetChangesets(String projectCode) in D:\code\languageforge-lexbox\backend\LexB
oxApi\Services\HgService.cs:line 67
at LexCore.Entities.Project.GetChangesets(IHgService hgService) in D:\code\languageforge-lexbox\backend\LexCor
e\Entities\Project.cs:line 17
at HotChocolate.Resolvers.Expressions.ExpressionHelper.AwaitTaskHelper[T](Task`1 task)
at HotChocolate.Types.Helpers.FieldMiddlewareCompiler.<>c__DisplayClass9_0.<<CreateResolverMiddleware>b__0>d.M
oveNext()
--- End of stack trace from previous location ---
at DataAnnotatedModelValidations.ValidatorMiddleware.InvokeAsync(IMiddlewareContext context)
at HotChocolate.Utilities.MiddlewareCompiler`1.ExpressionHelper.AwaitTaskHelper(Task task)
at HotChocolate.Execution.Processing.Tasks.ResolverTask.ExecuteResolverPipelineAsync(CancellationToken cancell
ationToken)
at HotChocolate.Execution.Processing.Tasks.ResolverTask.TryExecuteAsync(CancellationToken cancellationToken)
I realize now, I was actually running the api locally and everything else in k8s. If I run the api in k8s as well it works fine.
So, I assume what's happening here is that the actual repo is only on the local filesystem and so hg in k8s can't find it.
@hahn-kev @rmunn Is this something you've already thought through?
Yeah you're right. It'll work correctly if it's in k8s, but changes to the local file system won't make it into the volume. I tried to get skaffold to sync the files into the right place but I couldn't get it to work for some reason.
| gharchive/issue | 2023-07-25T07:29:55 | 2025-04-01T06:45:47.588264 | {
"authors": [
"hahn-kev",
"myieye"
],
"repo": "sillsdev/languageforge-lexbox",
"url": "https://github.com/sillsdev/languageforge-lexbox/issues/194",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2401747793 | Auto-parse all (many?) USFM resources to run test manually
Pull all USFM files from production Serval for Scripture Forge
Parse all of them and make sure that there is no crashing
Have it be "one click" or "two click" so that ti can be easily repeated.
Possibly a python script to pull all the resources and then a manual test in dotnet to run it.
@johnml1135 Can we close this now that the USFM issues are being tracked elswhere?
| gharchive/issue | 2024-07-10T21:09:59 | 2025-04-01T06:45:47.615454 | {
"authors": [
"Enkidu93",
"johnml1135"
],
"repo": "sillsdev/serval",
"url": "https://github.com/sillsdev/serval/issues/429",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
56123748 | Bug with my second import, please help
This is a project that involves RTL writing and Arabic Script
Undefined index: name line 86 /var/www/languageforge.org/htdocs/models/languageforge/lexicon/commands/LexProjectCommands.php
line , f->lex_import_lift(Array)
JsonRpcServer.php line 67, call_user_func_array(Array, Array)
api.php line 23, JsonRpcServer::handle(Sf, CI_Output)
Report this bug to the developers
Please copy error above and include screenshot
Please wait. Importing LIFT file...
Import/Export
LIFT Import
LIFT Export
Settings Import
Import a LIFT File
Drop a LIFT file here
or if you prefer...
Hi,
Thanks for trying to import your project on Language Forge and test it out. We recently made some changes to the import process making it more stable. Can you give the import another try and let us know how it goes?
Thanks,
Chris
for the Language Forge dev team
| gharchive/issue | 2015-01-31T12:41:14 | 2025-04-01T06:45:47.618836 | {
"authors": [
"azulforblue",
"megahirt"
],
"repo": "sillsdev/web-scriptureforge",
"url": "https://github.com/sillsdev/web-scriptureforge/issues/119",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
274392084 | Hide ConfigurationWindow when making mf change
so that the screen doesn't show Definition as meaning field before automatically changing to gloss
This change is
Reviewed 1 of 1 files at r1.
Review status: all files reviewed at latest revision, all discussions resolved.
Comments from Reviewable
| gharchive/pull-request | 2017-11-16T04:35:30 | 2025-04-01T06:45:47.623058 | {
"authors": [
"glasseyes",
"megahirt"
],
"repo": "sillsdev/wesay",
"url": "https://github.com/sillsdev/wesay/pull/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
235406115 | Chall12 accidental code blocks
@chall12 check your app reflection and your peer review. The spaces before your paragraphs (i.e. indentation) are interpreted as code blocks (that's an alternate way to write code beyond using backticks). Removing the extra spaces before each paragraph will fix this.
@eah13 Thanks for pointing that out! Edits are in pull requests #339 and #340
Fixed by #339 and #340
| gharchive/issue | 2017-06-13T00:54:30 | 2025-04-01T06:45:47.624386 | {
"authors": [
"chall12",
"eah13"
],
"repo": "silshack/summer2017",
"url": "https://github.com/silshack/summer2017/issues/324",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
108597896 | Subsites return 404 on every request
Main domain - company.example.com
Created a sub site with host as company.example.com/nz
Created a new page with parent as NZ site, but every page comes up with 404 error.
You'll need to create it as a subdomain (or separate domain), not as a separate path.
In other words - nz.company.example.com instead of company.example.com/nz
Thanks. Do you know of any module which creates subsites as sub directories just like wordpress environment does?
Thank you for your contribution here. I just wanted to let you know that we’re looking to improve our management of issues and pull requests, with a goal to providing clear direction on how Symbiote contributes to the open source community. As part of this ongoing effort:
issues and pull requests that have been open prior to December 2017 with no recent activity are being closed
this is included
if you can confirm this remains applicable for SS3.6+ please do open it again
remaining issues and pull requests will be
categorised
on a clear path to resolution (whatever the next step may be).
This ultimately brings us one step closer to our ambition; being able to prioritise current issues and contributions in a timely manner. If you have any questions, please see https://www.symbiote.com.au/contributing for further information.
| gharchive/issue | 2015-09-28T06:44:38 | 2025-04-01T06:45:47.636629 | {
"authors": [
"bbtdigitalkaran",
"nglasl",
"nyeholt"
],
"repo": "silverstripe-australia/silverstripe-multisites",
"url": "https://github.com/silverstripe-australia/silverstripe-multisites/issues/46",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2043624 | MINOR Fix test to use the non deprecated way of fetching database records
Dataobject#buildSQL is deprecated, use DataObject::get() instead, with the new data mapper there's no reason not to.
WOOO NEW ORM!
| gharchive/issue | 2011-10-25T10:37:03 | 2025-04-01T06:45:47.637653 | {
"authors": [
"sminnee",
"stojg"
],
"repo": "silverstripe/sapphire",
"url": "https://github.com/silverstripe/sapphire/issues/72",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1769256108 | SessionAuthenticationHandler doesn't cache member query
Affected Version
4.13+
Description
The SessionAuthenticationHandler is currently using
https://github.com/silverstripe/silverstripe-framework/blob/0c40cc9b3d9658ad868c560059c88255d2ee0674/src/Security/MemberAuthenticator/SessionAuthenticationHandler.php#L65
byID which doesn't cache the query. Which is a bit of a shame because the "current user" query is a very common one, eg:
Steps to Reproduce
Login and check the queries :-)
Fix
Use: $member = DataObject::get_by_id("SilverStripe\Security\Member", $id);
That's a good suggestion. Do you want to have a go at creating a PR?
already did : https://github.com/silverstripe/silverstripe-framework/pull/10834 :)
Merged - the auto-tagger should tag it once CI has finished running on it.
| gharchive/issue | 2023-06-22T09:02:59 | 2025-04-01T06:45:47.653993 | {
"authors": [
"GuySartorelli",
"lekoala",
"maxime-rainville"
],
"repo": "silverstripe/silverstripe-framework",
"url": "https://github.com/silverstripe/silverstripe-framework/issues/10833",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
461374858 | Wrong propagation of grid state in nested GridFields
Affected Version
4.4.1
Description
There is an issue when sorting and then navigating over nested GridFields. It seems as if the inner GridField will load the state of the parent's GridField which may contain invalid state data (e.g. Sort by a field not existing on the child). This will become a problem in GridFieldDetailForm_ItemRequest->getAdjacentRecordID.
Steps to Reproduce
Have two classes FooEntity and BarEntity where FooEntity has a many_many relation to BarEntity. FooEntity shall have some field FooField that does not exist on BarEntity.
Create a ModelAdmin for FooEntity and ensure that there is a GridField for the BarEntities inside the FooEntity EditForm.
Make FooField a summary field
Ensure that there are at least one object per class and a relation between them
Open the ModelAdmin for FooEntity and sort by FooField
Click on a FooEntity
Navigate to the GridField for the BarEntities and select one
See the mentioned error (note the sort field "FooField" in the grid state url)
And here a minimal test case:
https://gist.github.com/sb-relaxt-at/8da1f35f790a25ce4a46993867e947ab
https://gist.github.com/sb-relaxt-at/6e00c0277e86aca7dfcc41df082216cb
https://gist.github.com/sb-relaxt-at/897659e17962d0885d87b09f0307b9d7
I think this is a duplicate of https://github.com/silverstripe/silverstripe-admin/issues/887.
https://github.com/silverstripe/silverstripe-framework/pull/9072 should fix this.
I'll close this for now. If you look at the other issue and think it doesn't accurately describe your problem, I can reopen.
@maxime-rainville Thanks a lot, I must have overseen this issue, as the actual problem is inside framework and not admin. Looking forward for the bug to be fixed :-)
Hi,
I was just it by the same issue. I'm running on a project using v 4.4.3 and it does not seem fix. I'll also post on #9072 because I'm not sure why it's doing that.
| gharchive/issue | 2019-06-27T07:55:46 | 2025-04-01T06:45:47.660812 | {
"authors": [
"lekoala",
"maxime-rainville",
"sb-relaxt-at"
],
"repo": "silverstripe/silverstripe-framework",
"url": "https://github.com/silverstripe/silverstripe-framework/issues/9095",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
606425016 | Session::regenerateSessionId may cause warnings
Affected Version
4.5.3
Description
Recently found this in my logs in production:
error-log.WARNING: E_WARNING: session_regenerate_id(): Cannot regenerate session id - session is not active {"code":2,"message":"session_regenerate_id(): Cannot regenerate session id - session is not active","file":".../vendor/silverstripe/framework/src/Control/Session.php","line":652}
This happens when sending password reset links
Steps to Reproduce
Get a password reset link, open an incognito window, paste the reset link (in my case, it's hidden by a tracking link from the mail api, so it triggers a redirect) : you get the warning
It may well be due to the incognito window freshly opened or the redirection from the mail api. In any case, it shouldn't happen
Possible ways to solve it:
prepend @ in front of the call (like in vendor\silverstripe\framework\src\Security\MemberAuthenticator\SessionAuthenticationHandler.php)
check if session is active, something like this seem to work fine for me in my current hotfix
if (!headers_sent() && session_status() == PHP_SESSION_ACTIVE) {
session_regenerate_id(true);
}
Gah, we’ve already fixed this once (https://github.com/silverstripe/silverstripe-framework/issues/9259), obviously not properly :(
Fancy submitting a PR @lekoala?
@kinglozzer your changes are in the 4.5.3 tag (https://github.com/silverstripe/silverstripe-framework/blob/4.5.3/src/Control/Session.php) - so I guess it means there's still an edge case to be dealt with?
Yeah I think I fixed one trigger for this bug, but not the root cause
Fancy submitting a PR @lekoala?
PR done! https://github.com/silverstripe/silverstripe-framework/pull/9499
PR merged so this can be closed :D
| gharchive/issue | 2020-04-24T16:08:56 | 2025-04-01T06:45:47.666901 | {
"authors": [
"dhensby",
"kinglozzer",
"lekoala"
],
"repo": "silverstripe/silverstripe-framework",
"url": "https://github.com/silverstripe/silverstripe-framework/issues/9496",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2281934538 | TLN Update translations
Automated translations update generated using silverstripe/tx-translator
Replaced by https://github.com/silverstripe/silverstripe-totp-authenticator/pull/154
| gharchive/pull-request | 2024-05-06T23:07:09 | 2025-04-01T06:45:47.673063 | {
"authors": [
"GuySartorelli"
],
"repo": "silverstripe/silverstripe-totp-authenticator",
"url": "https://github.com/silverstripe/silverstripe-totp-authenticator/pull/153",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
325222839 | [sift] Could it be possible to support RGBA images?
Is that something easy or not?
On Tue, 22 May 2018 09:59:32 +0000 (UTC)
Valentin Valls notifications@github.com wrote:
Is that something easy or not?
We did it. (I do it instensively in my timelaps image alignment)
Basically it pre-processes the color image and renders it in black&white before using sift.
If it does not work, it is a bug.
--
Jérôme Kieffer
tel +33 476 882 445
Yes i see it is working well on RGB. But maybe you misunderstand me, i am talking about RGB+alpha.
On Tue, 22 May 2018 04:27:34 -0700
Valentin Valls notifications@github.com wrote:
Yes i see it is working well on RGB. But maybe you misunderstand me, i am talking about RGB+alpha.
Extending the code from RGB to RGBA shouldn't be difficult ...
| gharchive/issue | 2018-05-22T09:59:26 | 2025-04-01T06:45:47.681024 | {
"authors": [
"kif",
"vallsv"
],
"repo": "silx-kit/silx",
"url": "https://github.com/silx-kit/silx/issues/1841",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
281299324 | close #1405
bug which shows off in some dataset from Soleil
Looks fine. But send me your sample, i will maybe add a unittest.
| gharchive/pull-request | 2017-12-12T08:37:35 | 2025-04-01T06:45:47.681876 | {
"authors": [
"kif",
"vallsv"
],
"repo": "silx-kit/silx",
"url": "https://github.com/silx-kit/silx/pull/1406",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
915307428 | neon/st3q_u8: Wasm optimization
Autovectorizer is unable to do anything here, so implement it using a
bunch of shuffles.
Nice ☺
| gharchive/pull-request | 2021-06-08T17:52:34 | 2025-04-01T06:45:47.685035 | {
"authors": [
"nemequ",
"ngzhian"
],
"repo": "simd-everywhere/simde",
"url": "https://github.com/simd-everywhere/simde/pull/827",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2447267407 | fix alignment error in unit tests (gcc complains)
While validating #469 I saw that ub sanitizer with gcc complains about unaligned load.
This fixes the problem. clang does not complain.
Is there no CI job that runs gcc with sanitizers that has caught this?
I could not use the latest master because of the newly tweaked inline in 04b9b9ed1a698a161c71a6f901041735f19cd93b:
pauldreik@privat-kodning:~/code/delaktig/simdutf/tests/blah-gcc$ ninja && ctest -j100
[74/165] Building CXX object src/CMakeFiles/simdutf.dir/simdutf.cpp.o
FAILED: src/CMakeFiles/simdutf.dir/simdutf.cpp.o
/usr/lib/ccache/g++-12 -I/home/pauldreik/code/delaktig/simdutf/src -I/home/pauldreik/code/delaktig/simdutf/include -g -std=c++11 -Wall -Wextra -Weffc++ -Wfatal-errors -Wsign-compare -Wshadow -Wwrite-strings -Wpointer-arith -Winit-self -Wconversion -Wno-sign-conversion -Wunused-function -mno-avx256-split-unaligned-load -mno-avx256-split-unaligned-store -fsanitize=undefined -fno-sanitize-recover=all -MD -MT src/CMakeFiles/simdutf.dir/simdutf.cpp.o -MF src/CMakeFiles/simdutf.dir/simdutf.cpp.o.d -o src/CMakeFiles/simdutf.dir/simdutf.cpp.o -c /home/pauldreik/code/delaktig/simdutf/src/simdutf.cpp
In file included from /home/pauldreik/code/delaktig/simdutf/src/simdutf/haswell/simd.h:393,
from /home/pauldreik/code/delaktig/simdutf/src/simdutf/haswell.h:59,
from /home/pauldreik/code/delaktig/simdutf/src/implementation.cpp:27,
from /home/pauldreik/code/delaktig/simdutf/src/simdutf.cpp:4:
/home/pauldreik/code/delaktig/simdutf/src/simdutf/haswell/simd16-inl.h: In function ‘simdutf::haswell::{anonymous}::simd::simd16 simdutf::haswell::{anonymous}::simd::operator==(simd16, simd16)’:
/home/pauldreik/code/delaktig/simdutf/src/simdutf/haswell/simd16-inl.h:19:37: note: the ABI for passing parameters with 32-byte alignment has changed in GCC 4.6
19 | friend simdutf_really_inline Mask operator==(const simd16 lhs, const simd16 rhs) { return _mm256_cmpeq_epi16(lhs, rhs); }
| ^~~~~~~~
/home/pauldreik/code/delaktig/simdutf/src/simdutf/haswell/simd16-inl.h:19:126: warning: AVX vector return without AVX enabled changes the ABI [-Wpsabi]
19 | friend simdutf_really_inline Mask operator==(const simd16 lhs, const simd16 rhs) { return _mm256_cmpeq_epi16(lhs, rhs); }
| ^
In file included from /usr/lib/gcc/x86_64-linux-gnu/12/include/immintrin.h:47,
from /usr/lib/gcc/x86_64-linux-gnu/12/include/x86intrin.h:32,
from /home/pauldreik/code/delaktig/simdutf/src/simdutf/icelake/intrinsics.h:20,
from /home/pauldreik/code/delaktig/simdutf/src/simdutf/icelake.h:59,
from /home/pauldreik/code/delaktig/simdutf/src/implementation.cpp:26:
/usr/lib/gcc/x86_64-linux-gnu/12/include/avx2intrin.h: In function ‘simdutf::haswell::{anonymous}::simd::simd8 simdutf::haswell::{anonymous}::simd::operator==(simd8, simd8)’:
/usr/lib/gcc/x86_64-linux-gnu/12/include/avx2intrin.h:231:1: error: inlining failed in call to ‘always_inline’ ‘__m256i _mm256_cmpeq_epi8(__m256i, __m256i)’: target specific option mismatch
231 | _mm256_cmpeq_epi8 (__m256i __A, __m256i __B)
| ^~~~~~~~~~~~~~~~~
compilation terminated due to -Wfatal-errors.
[88/165] Building CXX object tests/CMakeFiles/base64_tests.dir/base64_tests.cpp.o
ninja: build stopped: subcommand failed.
Merging.
@pauldreik We do test in CI with GCC with sanitizers.
| gharchive/pull-request | 2024-08-04T19:29:18 | 2025-04-01T06:45:47.694301 | {
"authors": [
"lemire",
"pauldreik"
],
"repo": "simdutf/simdutf",
"url": "https://github.com/simdutf/simdutf/pull/474",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2278929134 | Share db with host Android app embedded flutter readonly/read write problems
We have an Android and iOS app that embeds flutter. Flutter uses drift to get some data from the app's db. We have disabled all migrations on drift side and recently changed db access from readWriteCreate to readOnly. (Mutex flag is null, and migration disabled here as well. closeUnderlyingOnClose set to true.)
I have created a repo where a simple subset of the above structure is used, and manifests a problem that might be related to what we are seeing in our production app.
If the Android app does lots of inserts, and some transactions in the background, then the embedded flutter sometimes gets SqliteException(11): while selecting from statement, database disk image is malformed, database disk image is malformed (code 11) in read write mode, and sometimes gets SqliteException(776): while selecting from statement, attempt to write a readonly database, attempt to write a readonly database (code 776) in read only mode.
I find the last result most surprising as the only query that is run is a select. I am not expecting any writes to be attempted from the flutter side.
As I understand how SQLite works, it would be valid to have drift access the SQlite file via the dart process in read only mode. (Even read write, if all the locks are working correctly between Android and Flutter. But this isn't as important for now, as our requirements are for Flutter to be read only for now.)
The Android app uses ORMLite, but connects to the SQLite db file using normal Android database APIs.
The emulator I've reliably reproduced this problem on is running Android API level 32
The project uses drift 2.17.0 and running flutter 3.19.6
Thanks for the report! I can reproduce the issue, but I'm afraid I don't really understand how this is happening either. Interestingly it works for me after re-opening the app and navigating to the Flutter fragment - so it's not an actual database corruption but rather appears to be corrupting an in-memory cache.
One workaround may be to use drift_sqflite instead of package:drift/native.dart and sqlite3_flutter_libs - you can replace the NativeDatabase with an SqfliteQueryExecutor. sqflite uses platform channels to run sqlite3 queries, so that avoids potential conflicts there.
I made a new branch on that repo that seems to confirm your suspicions. I can't now reproduce this problem on the example app. Thanks!
(I do notice that it is much slower, but I guess that is the cost of the workaround.)
Do you think it is worth posting the original problem on the SQLite forums? I am not expecting the Android process (That accesses the SQLite db file via the system SQLite service (and uses an older version of SQLite) to conflict with the one read/written by normal drift in this way.
Do you think it is worth posting the original problem on the SQLite forums?
I think the chance of this being an Android-specific issue are far higher. I know that Android is doing its own locking around sqlite3 databases when using the Java API, but I don't know if that means they're setting a flag to disable sqlite3 locks altogether. I wanted to check, but pragma compile_options; doesn't return any rows for me when running it on Android.
There has to be a proper solution here, but I could not find it. You can't be the first one sharing a sqlite3 database between Java and what's essentially an NDK thread accessing the same database via a direct C library for sqlite3.
I've seen an interesting thing: if you enable write-ahead logging on Android side, the problem goes away.
It is my understanding that the filesystem locking should allow for smooth access in this case between Android and drift. Could one of the two parties here be ignoring the locks? Or not setting them up correctly?
Patch snippet for above idea:
Index: app/src/main/java/com/skynamo/sqlitedbfun/ui/home/HomeFragment.kt
IDEA additional info:
Subsystem: com.intellij.openapi.diff.impl.patch.CharsetEP
<+>UTF-8
===================================================================
diff --git a/app/src/main/java/com/skynamo/sqlitedbfun/ui/home/HomeFragment.kt b/app/src/main/java/com/skynamo/sqlitedbfun/ui/home/HomeFragment.kt
--- a/app/src/main/java/com/skynamo/sqlitedbfun/ui/home/HomeFragment.kt (revision fdd421dfd08dda9613476f708192df3b57eb4a9d)
+++ b/app/src/main/java/com/skynamo/sqlitedbfun/ui/home/HomeFragment.kt (date 1715266975382)
@@ -64,6 +64,8 @@
private fun insertRandomUserData() {
if (_dao == null) {
_dao = DataItemDao(this.requireContext())
+ _dao?.databaseHelper?.setWriteAheadLoggingEnabled(true)
}
// Launch a coroutine in the background
Could one of the two parties here be ignoring the locks? Or not setting them up correctly?
Yeah that's what I'm thinking as well. It's possible that WAL mode just hides the symptoms (WAL mode still uses locks on the WAL file to avoid two concurrent writers - it's possible that this still has potential to break, but if one process is read-only it should be safe).
Still, drift/package:sqlite3 isn't disabling thread safety and as far as I know, Android isn't doing that either. So I still don't understand what's causing this.
I've been fighting with this problem few days. I have flutter app with drift (migrated from sqflite) and nice piece of native code (several widgets and Wear App). This pieces wants to connect to same db, using SQLiteDatabase from android.
sqflite worked with wal mode, it seems to be ok. But when I migrated to drift, I missed that there is other journal_mode by default. So I faced with bunch of issues with corrupted or busy sqlite db. I spent a few days on this, make it work somehow, but in some cases still have problems, but genrally it works.
Today I figure out, that database working in default journal_mode, not wal. Ok, I was set WAL and it starts crash again - corrupted db, busy mode or just nothing change in DB without any errors. I figured out, that reason of this - native side, I make small read query to DB, then close connection. When it happening - .wal and .shm files disappearing from filesystem and nothing commited to main DB. Even if drift connection still alive and continues insert/update db. This writes just do nothing.
For now I want to try advice from simolus3 and use drift_sqflite
sad
| gharchive/issue | 2024-05-04T10:45:50 | 2025-04-01T06:45:47.744552 | {
"authors": [
"diederikMrAwesome",
"simolus3",
"spChief"
],
"repo": "simolus3/drift",
"url": "https://github.com/simolus3/drift/issues/2990",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1576959953 | Remove hackyvariables from Postgres tests
I took a stab at improving the support for the Postgres dialect in the integration tests.
I fixed the variable binding test which were being skipped, but other issues remain which I will briefly mention.
One is the query amountOfGoodFriends which uses WHERE f.really_good_friends = 1.
That expression is invalid in postgres, as a bool is being compared with an int. Do you have any idea in mind with how to solve this?
The other is nested transaction support. Is there any reason why the Postgres implementation is not using the same logic as the SAVEPOINTs in the nested transactions for regular drift + sqlite?
I've encountered with https://github.com/isoos/postgresql-dart/issues/85 so I was looking for a way to handle nested transactions at the library level.
As for the PR, casting the SELECT expression in the test was the main fix, but for UInt8List it's not enough, as the output is "{1,2,3,..}" (with curly braces) instead of the bytes 1,2,3...
Thanks for the contribution! Recently, I helped with the postgres package to implement a new API that we can use to specify types without explicitly encoding them in the statement. I have done some work on integrating that in drift in the postgres3 branch. Since it will probably take a while to get that released though, I think this is a good improvement in the meantime.
For the amountOfGoodFriends query, we should just use TRUE instead of 1.
Is there any reason why the Postgres implementation is not using the same logic as the SAVEPOINTs in the nested transactions for regular drift + sqlite?
I think it was broken in some older versions of the postgres package, which encountered an illegal internal state when doing custom transaction management. We can probably remove the postgres-specific transaction implementation now.
@simolus3 I wasn't aware that the v3 API would support binding types without encoding, that's great to hear.
For the amountOfGoodFriends query, we should just use TRUE instead of 1.
I changed the 1 to TRUE to make it compatible.
I think it was broken in some older versions of the postgres package, which encountered an illegal internal state when doing custom transaction management. We can probably remove the postgres-specific transaction implementation now.
I changed the Transaction delegate to be the NoTransactionDelegate. Would that be the one that lets Drift use Savepoints internally?
Only 3 tests are failing now due to syntax error in the friendsOf query.
@simolus3 You are welcome! Thanks to you, your work on the drift ecosystem, superb as always 😄
| gharchive/pull-request | 2023-02-08T23:37:30 | 2025-04-01T06:45:47.751075 | {
"authors": [
"davidmartos96",
"simolus3"
],
"repo": "simolus3/drift",
"url": "https://github.com/simolus3/drift/pull/2308",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1723377399 | 🛑 Snow Boards NZ is down
In 9cb2a72, Snow Boards NZ (https://$SZ_SITE1/) was down:
HTTP code: 503
Response time: 1282 ms
Resolved: Snow Boards NZ is back up in bd16e62.
| gharchive/issue | 2023-05-24T07:42:44 | 2025-04-01T06:45:47.758567 | {
"authors": [
"simon2871"
],
"repo": "simon2871/fs-websitemonitor",
"url": "https://github.com/simon2871/fs-websitemonitor/issues/328",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2135762303 | 🛑 Snow Boards NZ is down
In 6976a82, Snow Boards NZ (https://$SZ_SITE1/) was down:
HTTP code: 503
Response time: 5609 ms
Resolved: Snow Boards NZ is back up in cca0f49 after 5 minutes.
| gharchive/issue | 2024-02-15T06:40:22 | 2025-04-01T06:45:47.760857 | {
"authors": [
"simon2871"
],
"repo": "simon2871/fs-websitemonitor",
"url": "https://github.com/simon2871/fs-websitemonitor/issues/939",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
565378436 | Rspec project Tic-Tact-toe
[x] Wrap specs for each method in describe block
[x] Instead of using before_each to define a variable for the test cases, use let.
[x] Use one it block for each expected behavior. Do not test multiple aspects in one it block.
[x] Add specs for all public methods you created in Tic Tac Toe project.
[x] Use existing repositories for Tic Tac Toe add specs and open new Pull Requests.
Pull Request under the review by Anna Muzykina (TSE)
Hi team!
Great job so far, however, there are a few things that are missing for this to be approved.
All the public methods of the classes implemented in project OOP should be tested
All If specs should cover positive and negative scenarios
Comments: It’s better to delete commented code. If you need to review a previous version of your code, use git history
Make changes and submit another code review request.
If you have any questions, let me know on slack (@Anna-Muzykina) and I will be glad to clarify with you.
Happy coding 😄
Status Require: changes🛑
🙋♂️@Certil
This what I get after running Rspec🙂
👉
Kindly rectify this error and submit another review
Happy coding
Addo Forison
Hi , It's say in the Readme run bundle install to install required Gem.
Status Require: changesstop_sign
raising_hand_man@Certil
This what I get after running Rspecslightly_smiling_face
point_right
Kindly rectify this error and submit another review
Happy coding
Addo Forison
Hi , It's say in the Readme run bundle install to install required Gem.
sorry for earlier
Status: Approved💚
sorry for earlier😔,
Your project is good💪
Go ahead and merge
No problem it could happen to anyone. Thank you for reviewing our project out of business
time.
| gharchive/pull-request | 2020-02-14T14:58:32 | 2025-04-01T06:45:47.793789 | {
"authors": [
"Anna-Myzukina",
"Forison",
"certilremy"
],
"repo": "simonpeterdebbarma/tic-tac-toe",
"url": "https://github.com/simonpeterdebbarma/tic-tac-toe/pull/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1000208367 | [app-state] more informative error message when assigning to an undefined state
When assigning to a state that is undefined, (i.e. store('mystate').state() === undefined), the library produces an error that is confusing and does not actually tell me what I'm doing wrong.
i.e.
store('mystate').assign(undefined)
store('mystate').assign({age: 30, name: 'joe'})
produces a console error saying:
preview-bc361a02a95aae7c3363b.js:2 TypeError: Cannot read property 'age' of undefined
at eval (store.ts:73)
at eval (every.ts:27)
Desired behaviour
A message saying "cannot assign to undefined property mystate"
(and perhaps include the application code line number in the stack trace?)
Stackblitz link:
https://stackblitz.com/edit/typescript-qg4jq7
I like your profile pic :)
Thanks for the suggestion!
This will come with version 13
| gharchive/issue | 2021-09-19T04:52:09 | 2025-04-01T06:45:47.808989 | {
"authors": [
"RaphiStein",
"ersimont"
],
"repo": "simontonsoftware/s-libs",
"url": "https://github.com/simontonsoftware/s-libs/issues/53",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1078969440 | [app-state]: Feature request: PersistedStore
It's a common pattern to combine a RootStore, Persistence and MigrationManager. It would be nice to reduce a little boilerplate. Something like:
export class PersistedStore<T extends VersionedObject> extends RootStore<T> {
constructor(
persistenceKey: string,
defaultState: T,
migrator: MigrationManager<T>,
) {
const persistence = new Persistence<T>(persistenceKey);
super(migrator.run(persistence, defaultState));
this.$.subscribe((state) => persistence.put(state));
}
}
Added. Will be available in the next release.
| gharchive/issue | 2021-12-13T20:24:21 | 2025-04-01T06:45:47.810663 | {
"authors": [
"eric-simonton-sama",
"ersimont"
],
"repo": "simontonsoftware/s-libs",
"url": "https://github.com/simontonsoftware/s-libs/issues/67",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2467261532 | Tests for the fix from #27 and #21
This looks harmless enough that I'll merge it without tests.
Originally posted by @simonw in https://github.com/simonw/asgi-csrf/issues/27#issuecomment-2290444922
It would be great to add tests for this - need steps to reproduce for this issue first:
#21
Oops, the lack of tests took us to <100% coverage: https://github.com/simonw/asgi-csrf/actions/runs/10398021931/job/28794689027
Here's that coverage report:
Surrounding code: https://github.com/simonw/asgi-csrf/blob/a18f7d0c9ad307709578f20b94b7fbbea112e857/asgi_csrf.py#L262-L305
I think this might occur on an empty POST form submission.
I spent a while on this and still couldn't recreate it, so I've dropped the coverage threshold to 99% so I can move on and work on other things.
| gharchive/issue | 2024-08-15T02:55:53 | 2025-04-01T06:45:47.816244 | {
"authors": [
"simonw"
],
"repo": "simonw/asgi-csrf",
"url": "https://github.com/simonw/asgi-csrf/issues/31",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
952327740 | --scope teamname option
To enable publishing to a different team than the default.
datasette publish vercel blah.db --project hello --team myteam
I'm going to test this with https://github.com/simonw/til
That worked: https://github.com/simonw/til/runs/3155919032
| gharchive/issue | 2021-07-25T18:14:48 | 2025-04-01T06:45:47.818025 | {
"authors": [
"simonw"
],
"repo": "simonw/datasette-publish-vercel",
"url": "https://github.com/simonw/datasette-publish-vercel/issues/47",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
803929694 | Try profiling Datasette using scalene
https://github.com/emeryberger/scalene looks like an interesting profiling tool.
https://github.com/emeryberger/scalene/issues/110 reports a "received signal SIGSEGV" error that was fixed by upgrading to the latest Scalene version, but I'm running that already.
| gharchive/issue | 2021-02-08T20:37:06 | 2025-04-01T06:45:47.821705 | {
"authors": [
"simonw"
],
"repo": "simonw/datasette",
"url": "https://github.com/simonw/datasette/issues/1219",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
490737169 | Version 50+ having trouble picking up initial notifications.
There was on scrobble in the cache, which got pushed to last.fm, but it can't pick up any more. I've uninstaller, reset caches and removed local data, but it's still not noticing that anything new has been played (to push to the now playing api) or has been played.
Well that's really not good.
This is my apk or the playstore version?
Double checked and it seems to be both.
Migjt be something in the app detection code; I can't get it to pick up new apps either.
Strange.
Have you tried Exiting the app? I added an exit button on the initial screen.
Sometimes this helps.
This is really bad news.
I did not change any of the code there.
Ok I recreated your bug.
There is an app clash problem upon initial install I believe.
If I play apps that previously used Broadcast intents it clashes.
Do you happen to know if all the apps you have been trying use broadcast intents in the old way?
I looked at Androids example mediamanager external player app and found out the real issue.
Thanks so much for pointing this issue out.
#508
Moving conversation from #507 to here like I was told earlier.
After trying 1.6.2 using multiple media players (Youtube, Play Music, GMMP) SLS still doesn't pick up any of the apps in the "Enabled Apps" section, nor can I get it to scrobble. Not sure if others are still having issues on 1.6.2
Also a potentially unrelated Issue that might need a separate thread, the stats from the older version (1.5.7 in my case) are still missing in the last several updates (1.6.0-1.6.2 etc), yet if I revert to version 1.5.7 the Stats show up fine and are still there
Yes, stats are not typically kept through versions.
I will add that to my todo list.
However, version 1.6.2 should pick up new music players now.
Did you enable the Notification Listener AKA Notification Access permissions?
Ah it seems I was mistaken, I went and made sure I had that permission enabled and it was off/not allowed. After allowing that Notification permission it seems as though at least the Now Playing portion is working, but songs aren't scrobbling once finished.
I was a bit confused because I saw we were now allowed to skip the permission if we didn't want to enable them all. I was under the assumption that the Notification Access was only needed if you used the Active App Notification option
| gharchive/issue | 2019-09-08T10:24:40 | 2025-04-01T06:45:47.865104 | {
"authors": [
"WolfFhang",
"a93h",
"djryan"
],
"repo": "simple-last-fm-scrobbler/sls",
"url": "https://github.com/simple-last-fm-scrobbler/sls/issues/503",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2171522864 | 🛑 SimpleLogin Landing Page is down
In 3f245b0, SimpleLogin Landing Page (https://simplelogin.io) was down:
HTTP code: 404
Response time: 636 ms
Resolved: SimpleLogin Landing Page is back up in cf198bf after 6 minutes.
| gharchive/issue | 2024-03-06T13:31:44 | 2025-04-01T06:45:47.868121 | {
"authors": [
"nguyenkims"
],
"repo": "simple-login/uptime",
"url": "https://github.com/simple-login/uptime/issues/83",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
59422401 | Add env_datetime to library
What is the problem / feature ?
env_datetime is sorta nice to have.
How did it get fixed / implemented ?
Added an env_datetime. Provides two supported formats, timestamp and iso8601
I specifically chose not to use the dateutils library because I don't think this library should have any dependencies since it's such a low level tool.
How can someone test / see it ?
Look at the README.
Here is a cute animal picture for your troubles...
Stated above but...:
I specifically chose not to use the dateutils library because I don't think this library should have any dependencies since it's such a low level tool.
:+1:
@cevaris
I updated this. PR.
Yeah, if it is not a datetime is not formatted in timestamp or iso8601, the developer still has the ability to use env_string and parse it manually themselves.
Yeah, I would punt. I can se iso8601 formatted dates in ENV variables, but
timestamps should be rare, if not convertible to iso8601 format.
On Wed, Mar 11, 2015 at 8:09 PM, Piper Merriam notifications@github.com
wrote:
Merged #5 https://github.com/simpleenergy/env-excavator/pull/5.
—
Reply to this email directly or view it on GitHub
https://github.com/simpleenergy/env-excavator/pull/5#event-250598625.
--
Warm Regards,
Julian Cevaris
Blogspot http://draft-new-religion.blogspot.com/ - Google+
https://plus.google.com/111474132801017393519/posts - Facebook
http://www.facebook.com/profile.php?id=100003308886461 - Twitter
http://twitter.com/#!/cevaris
| gharchive/pull-request | 2015-03-01T23:02:15 | 2025-04-01T06:45:47.876542 | {
"authors": [
"Thanavath",
"cevaris",
"pipermerriam"
],
"repo": "simpleenergy/env-excavator",
"url": "https://github.com/simpleenergy/env-excavator/pull/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2325212965 | deprecate old csi API endpoints
Tested the changes by running e2e tests.
Will run 5 of 5 specs
------------------------------
• [SLOW TEST] [61.578 seconds]
CSI Driver tests Control Plane: delete second lvol while the first lvol has IO running if a node has lvol with IO running, adding and deleting an new lvol from the same node should work
/home/ubuntu/actions-runner/_work/simplyBlockDeploy/simplyBlockDeploy/spdk-csi/e2e/controlplane.go:26
------------------------------
•
------------------------------
• [SLOW TEST] [50.774 seconds]
SPDKCSI-NVMEOF Test SPDK CSI Dynamic Volume Provisioning Test the flow for Dynamic volume provisioning
/home/ubuntu/actions-runner/_work/simplyBlockDeploy/simplyBlockDeploy/spdk-csi/e2e/nvmeof.go:30
------------------------------
• [SLOW TEST] [122.170 seconds]
SPDKCSI-NVMEOF Test SPDK CSI Dynamic Volume Provisioning Test the flow for Caching nodes
/home/ubuntu/actions-runner/_work/simplyBlockDeploy/simplyBlockDeploy/spdk-csi/e2e/nvmeof.go:51
------------------------------
• [SLOW TEST] [80.269 seconds]
SPDKCSI-NVMEOF Test SPDK CSI Dynamic Volume Provisioning Test multiple PVCs
/home/ubuntu/actions-runner/_work/simplyBlockDeploy/simplyBlockDeploy/spdk-csi/e2e/nvmeof.go:79
------------------------------
Ran 5 of 5 Specs in 315.185 seconds
SUCCESS! -- 5 Passed | 0 Failed | 0 Pending | 0 Skipped
--- PASS: TestE2E (315.19s)
PASS
ok github.com/spdk/spdk-csi/e2e 316.544s
We can't merge yet until @Hamdy-khader merges this PR if not the route will break in sbcli-dev
we still need to deprecate the /csi calls in https://github.com/simplyblock-io/spdk-csi/blob/317d9799f9c3773760089fa1aa3e0796ea4adf1d/pkg/util/nvmf.go#L153-L173
we still need to deprecate the /csi calls in
https://github.com/simplyblock-io/spdk-csi/blob/317d9799f9c3773760089fa1aa3e0796ea4adf1d/pkg/util/nvmf.go#L153-L173
I missed this.
| gharchive/pull-request | 2024-05-30T09:38:41 | 2025-04-01T06:45:47.905261 | {
"authors": [
"boddumanohar",
"geoffrey1330"
],
"repo": "simplyblock-io/spdk-csi",
"url": "https://github.com/simplyblock-io/spdk-csi/pull/43",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
145572068 | Chinese Dream
By ZYMK07
And I suggest you should give more flexbility of the proletarian politics especially after the reform and open policy(could be the decision for all proletarian countries in the game). The socialism is quiet different from the Soviet after the reform and the ruling party also begin to political reform.
By Simulla
Agree on more flexibility of the proletarian politics. Plan to have "Chinese Dream" to counter "western democracy", "American Dream", if China becomes rank 1.
There is already market communism in the mod that reflect flexibility of the proletarian politics after the reform and open policy. Since this is largely assassinated with modern development in China I have "Chinese Dream" event if China becomes rank 1.
As part of this issue some sensitive material has been removed. Overall harmonizing the mod with the "Secunde Directive".
| gharchive/issue | 2016-04-04T03:14:55 | 2025-04-01T06:45:47.941795 | {
"authors": [
"simsulla"
],
"repo": "simsulla/nwomod",
"url": "https://github.com/simsulla/nwomod/issues/304",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
108007777 | display missing tooltips on hover
Issue #121
Haven't found how to prevent build failed.
Look at the travis output: https://travis-ci.org/sindresorhus/atom-jshint/builds/81857827
It's about code style.
Use
const target = editorElement.shadowRoot.querySelectorAll(`.line-number[data-buffer-row="${row}"]`);
```
instead of
```ES6
const target = editorElement.shadowRoot.querySelectorAll('.line-number[data-buffer-row="' + row + '"]');
```
Even the topic closed, I didn't understand how to solve the problem.
Problem solved with the update to the 1.8.3 version
| gharchive/pull-request | 2015-09-23T21:13:49 | 2025-04-01T06:45:47.990428 | {
"authors": [
"Charlie-Y",
"helloncanella",
"sindresorhus",
"thisninja"
],
"repo": "sindresorhus/atom-jshint",
"url": "https://github.com/sindresorhus/atom-jshint/pull/122",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
118760298 | Hide our own error stack on IPC messaged errors
From: https://github.com/sindresorhus/ava/issues/245#issuecomment-158339191
Our error stack is pretty useless when an error comes from the forked process.
Agreed.
I think the best solution is to stop rejecting the promise in that case. Rejection has the side affect of bailing out of Promise.all early, so other test files may not finish. We should resolve with an failure result, that will simply be summarized in the output. We should save rejection for unexpected/unrecoverable problems.
We should resolve with an failure result, that will simply be summarized in the output. We should save rejection for unexpected/unrecoverable problems.
:+1:
Pretty sure this has been addressed in various ways. Closing.
| gharchive/issue | 2015-11-25T04:43:27 | 2025-04-01T06:45:47.993065 | {
"authors": [
"jamestalmage",
"novemberborn",
"sindresorhus"
],
"repo": "sindresorhus/ava",
"url": "https://github.com/sindresorhus/ava/issues/269",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
138852019 | Add Reference Tools
https://github.com/willhoag/awesome-reference-tools
The list is not very comprehensive. Awesome lists are hard work. Put some more work into it. Take a look at some other lists and see how they do it. E.g. https://github.com/sindresorhus/awesome-electron https://github.com/sindresorhus/awesome-npm
Would be happy to reconsider in a month when the list has gotten some more work.
It's also missing a license. Please read the awesome manifest.
Yep, sounds good. I'll be sure to invest some more time into it and will come back when it is ready. And, add a license. Thanks!
| gharchive/pull-request | 2016-03-07T00:02:59 | 2025-04-01T06:45:47.995869 | {
"authors": [
"sindresorhus",
"willhoag"
],
"repo": "sindresorhus/awesome",
"url": "https://github.com/sindresorhus/awesome/pull/563",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
360529732 | Regressions with Github Enterprise
Removal of functionality that public Github has added but has yet to be added to Github Enterprise can cause regressions for some users.
For example, the functionality provided here
https://blog.github.com/changelog/2018-08-17-collapse-all-diffs-in-a-pull-request-at-once/
which was removed with https://github.com/sindresorhus/refined-github/pull/1524 doesn't yet exist in the enterprise version.
Not sure if this is feasible but maybe for things that get added to public GH, we should disable them by default with an option for users to opt in, instead of removing them entirely?
We try not to break GitHub Enterprise and we accept Enterprise-only bug fixes, but we’re not going to hold back doing non-breaking changes like this. If you need more control, you can clone the repo and use the extension as a local extension. That way you can control when to update.
| gharchive/issue | 2018-09-15T10:53:52 | 2025-04-01T06:45:48.077034 | {
"authors": [
"ashrayjain",
"sindresorhus"
],
"repo": "sindresorhus/refined-github",
"url": "https://github.com/sindresorhus/refined-github/issues/1530",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
416247938 | Refined GitHub version history: compatibility with versions of Firefox
Somewhere between (not within) these two records:
https://web.archive.org/web/20181002162604/https://addons.mozilla.org/en-US/firefox/addon/refined-github-/versions/ (ending at 18.9.28.1714)
https://addons.mozilla.org/addon/refined-github-/versions/ (no longer allowing more than 25 versions)
– there'll be the last version that was compatible with Firefox 56 (or 55.0 and later).
Please, can you tell which version? Plus a link to the .xpi, if you can.
Thanks
(Sorry, the bug label was not intended; I can't remove it.)
What version of Firefox do you need exactly? And why are you on an old version of Firefox?
Thanks for identifying the commits.
I have Waterfox 56.2.7.1. Waterfox 56.0 was based on Firefox 56.0.
Re: https://redd.it/aw8rqm whilst I happily use slightly outdated 19.1.31.1822 (for Firefox 61.0 and later, not expecting full compatibility with Waterfox 56.x), other users of Waterfox might prefer the most recent version of the extension that was compatible with Firefox 56.
I'll raise a question in the Mozilla area. If we can be given a real-world example of how to use the API – resulting in an AMO link to the required .xpi – then future situations such as might become less complicated for people (like me) who are not developers …
There's no other way to get the xpi, we don't save each version. You can find the URLs on Wayback. To download them, strip everything before the second https in the URL:
https://web.archive.org/web/20181002162604/https://addons.mozilla.org/firefox/downloads/file/1092770/refined_github-18.9.28.1714-an+fx.xpi?src=version-history
Becomes:
https://addons.mozilla.org/firefox/downloads/file/1092770/refined_github-18.9.28.1714-an+fx.xpi?src=version-history
The commit I mentioned was on version 18.12.30
Also: https://blog.mozilla.org/firefox/firefox-64-default-64-bit-windows/
I don't suggest using a browser based on Firefox 56 (released on September 28, 2017)
Thanks, Waterfox is patched for MFSAs and so on.
You can find the URLs on Wayback.
Unfortunately not for what's required here, that's why I began the opening post with a Wayback Machine link.
There's a slim possibility that someone captured a non- en-US/ variation of the required page at the required point in time, but this is very slim. https://web.archive.org/web/*/https://addons.mozilla.org/en-GB/firefox/addon/refined-github-/versions/ captured only once (by me), and so on.
I experimented with https://addons.mozilla.org/firefox/downloads/file/1092770/refined_github-18.12.30-an+fx.xpi but this results in version 18.9.28.1714 (not version 18.12.30).
Lateral thinking, I installed an overly outdated version then looked to see what might be found with a check for updates.
Found: 18.12.22.1803
I can't tell the URL of the .xpi, but it's one way to get close to 18.12.30.
| gharchive/issue | 2019-03-01T19:32:31 | 2025-04-01T06:45:48.087787 | {
"authors": [
"bfred-it",
"grahamperrin"
],
"repo": "sindresorhus/refined-github",
"url": "https://github.com/sindresorhus/refined-github/issues/1819",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
540846142 | Status filters on PR list page display JS code
You need a repo that has some checks
Go to the PR list page (i.e. https://github.com/sindresorhus/refined-github/pulls?q=is%3Apr+is%3Aopen+sort%3Aupdated-desc)
Click on the Status filter and see that it shows JS code in the dropdown items, i.e.:
function() {
return dom_chef_default.a.createElement("svg", {
class: "octicon octicon-check",
xmlns: "http://www.w3.org/2000/svg",
width: "12",
height: "16",
viewBox: "0 0 12 16"
}, dom_chef_default.a.createElement("path", {
"fill-rule": "evenodd",
d: "M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5L12 5z"
}));
}
Thank you for the report!
| gharchive/issue | 2019-12-20T08:13:42 | 2025-04-01T06:45:48.090966 | {
"authors": [
"arnaudvalle",
"fregante"
],
"repo": "sindresorhus/refined-github",
"url": "https://github.com/sindresorhus/refined-github/issues/2646",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
775826066 | Fix stop-pjax-loading-with-esc bug that breaks profile timeline
LINKED ISSUES:
Closes #3789
TEST URLS:
https://github.com/dnicolson
SCREENSHOT:
Is this code no longer needed at all? 🎉
| gharchive/pull-request | 2020-12-29T10:45:35 | 2025-04-01T06:45:48.093138 | {
"authors": [
"artusm",
"fregante"
],
"repo": "sindresorhus/refined-github",
"url": "https://github.com/sindresorhus/refined-github/pull/3856",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
236653620 | Optimize and group PNGs
With ImageOptim.app
| gharchive/pull-request | 2017-06-17T10:46:36 | 2025-04-01T06:45:48.093842 | {
"authors": [
"algern9n",
"bfred-it"
],
"repo": "sindresorhus/refined-github",
"url": "https://github.com/sindresorhus/refined-github/pull/500",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
458227341 | Documentation from Companies
Doc Geral has a column for the documentation (whether it has been requested, received, etc).
Should we introduce this in the DB?
Maybe this should go to partners?
Goes to partners
| gharchive/issue | 2019-06-19T21:07:48 | 2025-04-01T06:45:48.108137 | {
"authors": [
"BSDinis"
],
"repo": "sinfo/deck2",
"url": "https://github.com/sinfo/deck2/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2260978148 | how i can exclude device_info_plus
Is there an existing issue for this?
[X] I have searched the existing issues
The question
how i can exclude device_info_plus
Unfortunately there is no way since the editor depends on it, the solution would be that we extract all the plugins code into a separated package that implements an interface that we create in the core package and requiring to implement the interface when using the core package or without nothing when using the new package that include everything of the core package + implementing the interface with the plugins code
so users can define their own solutions or use completely different package but this will require the user/developer some work to get it done
Issue fixed in 10.5.15-dev.0.
The device_info_plus plugin has been removed and is no longer a dependency of flutter_quill
| gharchive/issue | 2024-04-24T10:38:54 | 2025-04-01T06:45:48.111114 | {
"authors": [
"EchoEllet",
"ellet0",
"samih93"
],
"repo": "singerdmx/flutter-quill",
"url": "https://github.com/singerdmx/flutter-quill/issues/1833",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1309179122 | [Mobile]
My issue is about [Mobile]
I have tried running example directory successfully before creating an issue here.
Please note that we are using latest flutter version in stable channel on branch master. If you are using beta or master channel, or you are not using latest flutter version in stable channel, you may experience error.
Hey man,i found a bug when i am inputting chinese,the style is not effective unless input a space first.
There are some differences between chinese and the others.there is a example for you,i wish you will understand.
eg: words1 => ni hao
words2 => 你好
the Editor always show words1 when somebody inputing,it contains the style you have setted.but something wrong when you comfirm,the action means words1 change to words2.the style will miss. I dont know how to solve it.
If I miss something,could you pls tell me.
It would help a lot of you can share a screen capture. Also please consider updating the title. It's not helpful for setting prios for devs that have time to contribute. Cheers!
It would help a lot of you can share a screen capture. Also please consider updating the title. It's not helpful for setting prios for devs that have time to contribute. Cheers!
thanks,bro.i change the title.i will take a photo or video for you
It would help a lot of you can share a screen capture. Also please consider updating the title. It's not helpful for setting prios for devs that have time to contribute. Cheers!
thanks,bro.i change the title.i will take a photo or video for you
hi,bro.i provide a video to show the bug.i hope it is helpful--https://www.clzg.cn/article/384286.html
| gharchive/issue | 2022-07-19T08:37:54 | 2025-04-01T06:45:48.115907 | {
"authors": [
"adrian-moisa",
"fuwuyuan"
],
"repo": "singerdmx/flutter-quill",
"url": "https://github.com/singerdmx/flutter-quill/issues/872",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1955158448 | Tip: How To Loop Through Slides and/or Programatically Replace Text
It took me hours to figure this out so thought I'd share a quick workaround on how to:
i) loop through all the slides in a presentation
ii) replace all instances of the text you want to replace, without needed to know/hardcode the selector names
This won't work for text in tables though, if anyone has a tip on how to do that, I'd be happy to try.
let variableData = [{variable: "First Name", value: "John"}, {variable: "Last Name", value: "Doe"}]
let textReplacements = [];
let templateName = 'test';
variableData.forEach(async variable => {
textReplacements.push({
replace: variable.variable,
by: {
text: variable.value,
},
});
});
// Load Pptx Automizer
const automizer = new Automizer({
templateDir: templateDir,
outputDir: outputDir,
removeExistingSlides: true,
});
let pres = automizer.loadRoot(rootTemplate).load(extraTemplate, templateName);
// Count the total number of slides
const creationIds = await pres.setCreationIds();
let totalSlides = 0;
creationIds.forEach((template) => {
totalSlides += template.slides.length;
});
console.log(`Processing ${totalSlides} slides!`);
// Loop through all the slides
for (let i = 1; i <= totalSlides; i++) {
pres.addSlide(templateName, i, async slide => {
let placeholders = await slide.getAllTextElementIds();
// Replace text for each placeholder programatically
placeholders.forEach(async placeholder => {
slide.modifyElement(
placeholder,
modify.replaceText(textReplacements),
);
});
});
}
// Write the output file
pres.write('Output.pptx').then((summary) => {
console.log(summary);
});
Great! Thanks a lot for sharing! :smiley:
You can make use of the new pres.getInfo() that came yesterday with v0.4.1 to make the code even shorter:
// ...
let pres = automizer.loadRoot(rootTemplate).load(extraTemplate, templateName);
// getInfo() is a tiny wrapper around setCreationIds():
const presInfo = await pres.getInfo();
// Count the total number of slides
const totalSlides = presInfo.slidesByTemplate(templateName).length;
console.log(`Processing ${totalSlides} slides!`);
// Loop through all the slides
// ...
Best regards and keep on sharing!
How to deal with only one template (the root template)?
I found a strange problem where some fields cannot be replaced, such as: installedCapacity. Can someone help me try it out?
I found a strange problem where some fields cannot be replaced, such as: installedCapacity. Can someone help me try it out?
I found a strange problem where some fields cannot be replaced, such as: installedCapacity. Can someone help me try it out?
It took me hours to figure this out so thought I'd share a quick workaround on how to: i) loop through all the slides in a presentation ii) replace all instances of the text you want to replace, without needed to know/hardcode the selector names
This won't work for text in tables though, if anyone has a tip on how to do that, I'd be happy to try.
let variableData = [{variable: "First Name", value: "John"}, {variable: "Last Name", value: "Doe"}]
let textReplacements = [];
let templateName = 'test';
variableData.forEach(async variable => {
textReplacements.push({
replace: variable.variable,
by: {
text: variable.value,
},
});
});
// Load Pptx Automizer
const automizer = new Automizer({
templateDir: templateDir,
outputDir: outputDir,
removeExistingSlides: true,
});
let pres = automizer.loadRoot(rootTemplate).load(extraTemplate, templateName);
// Count the total number of slides
const creationIds = await pres.setCreationIds();
let totalSlides = 0;
creationIds.forEach((template) => {
totalSlides += template.slides.length;
});
console.log(`Processing ${totalSlides} slides!`);
// Loop through all the slides
for (let i = 1; i <= totalSlides; i++) {
pres.addSlide(templateName, i, async slide => {
let placeholders = await slide.getAllTextElementIds();
// Replace text for each placeholder programatically
placeholders.forEach(async placeholder => {
slide.modifyElement(
placeholder,
modify.replaceText(textReplacements),
);
});
});
}
// Write the output file
pres.write('Output.pptx').then((summary) => {
console.log(summary);
});
I found a strange problem where some fields cannot be replaced, such as: installedCapacity. Can someone help me try it out?
Did you try to disable spell checking for the template field?
@singerla I'm working on a task which replaces the text present in the pptx of a particular slide. Also all functionality has been implemented in python using python-pptx lib. But I'm struggle at replacing text present in SmartArt, Charts etc. Does this pptx-automizer supports modifying SmartArt, Charts etc?
@b1ues Did disabling spell check solve your problem? Feel free to continue on #82 otherwise.
| gharchive/issue | 2023-10-21T00:27:16 | 2025-04-01T06:45:48.123022 | {
"authors": [
"HarshalSangme",
"Suniron",
"b1ues",
"singerla",
"vertex-admin"
],
"repo": "singerla/pptx-automizer",
"url": "https://github.com/singerla/pptx-automizer/issues/72",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2531375600 | [python] Remove pyarrow pin on MacOS
Long-awaited work coming to fruition:
https://github.com/single-cell-data/TileDB-SOMA/issues/1849
https://github.com/single-cell-data/TileDB-SOMA/pull/2692
https://github.com/single-cell-data/TileDB-SOMA/pull/2734
https://github.com/TileDB-Inc/TileDB/pull/5223 went into core 2.26, making this possible
See also https://github.com/single-cell-data/TileDB-SOMA/issues/2999 [sc-53002]
@teo-tsirpanis has approved; not sure what happened to his approval but it's past EOD in his timezone
I don't have write permissions to this repository.
Ah, thanks @teo-tsirpanis ! :)
| gharchive/pull-request | 2024-09-17T14:50:02 | 2025-04-01T06:45:48.127543 | {
"authors": [
"johnkerl",
"teo-tsirpanis"
],
"repo": "single-cell-data/TileDB-SOMA",
"url": "https://github.com/single-cell-data/TileDB-SOMA/pull/3004",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1281920550 | feat(streaming): remove source limit for concurrent ckpt
Signed-off-by: Alex Chi iskyzh@gmail.com
I hereby agree to the terms of the Singularity Data, Inc. Contributor License Agreement.
What's changed and what's your intention?
Should be merged after concurrent checkpoint. Remove source limit.
Checklist
[x] I have written necessary docs and comments
[x] I have added necessary unit tests and integration tests
[x] All checks passed in ./risedev check (or alias, ./risedev c)
Refer to a related PR or issue link (optional)
close https://github.com/singularity-data/risingwave/issues/3403
is the backpressure mechanism not necessary now?
@skyzh #3415 is merged. We can resolve the conflict and merge this PR now.
| gharchive/pull-request | 2022-06-23T07:07:44 | 2025-04-01T06:45:48.134608 | {
"authors": [
"hzxa21",
"skyzh",
"tabVersion"
],
"repo": "singularity-data/risingwave",
"url": "https://github.com/singularity-data/risingwave/pull/3424",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2077626085 | Read stadiamap key from file
After meeting with Hubzero's Steve & Pascal, the solution for securely storing the Google Stadiamap API key is a new feature in the hub which allows storing sensitive information that can only be access by the tool at runtime.
To implement this, we need our app to be able to read the key from a file and then call register_stadiamaps(key="SEE SLACK", write=FALSE).
For more info, see issue #67
reopening this issue with additional info from Pascal @ Hubzero:
Hi Mona! On uccommunityhub, I set dummy example secrets for a regular development workspace tool and for mkde. There’s a user, password, host, and “var1” defined, one line for each. The paths are respectively /secrets/workspace.txt and /secrets/mkde.txt. Let me know which real secrets you want there
so need to update the code with the new file location...
Pascal's Jira ticket @ Hubzero...
| gharchive/issue | 2024-01-11T21:31:31 | 2025-04-01T06:45:48.155104 | {
"authors": [
"monaw"
],
"repo": "sinkovit/Spatial-ecology",
"url": "https://github.com/sinkovit/Spatial-ecology/issues/78",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1358281333 | multivariate data on Tulog.ipynb
Orion version: 0.3.2
Python version:3.7
Operating System: linux
Description
Hi I tried using a multivariate data ('multivariate/S-1') which has 25 signals on Tulog.ipynb
The Orion API at part 2 is able to fit the data.
However at the Pipeline Training and Detection Section using primitives,
from model import hyperparameters
from orion.primitives.tadgan import TadGAN
hyperparameters["epochs"] = 5
hyperparameters["input_shape"] = (100, 25) # based on the window size
hyperparameters["optimizer"] = "keras.optimizers.Adam"
hyperparameters["learning_rate"] = 0.0005
hyperparameters["latent_dim"] = 20
hyperparameters["batch_size"] = 64
tgan = TadGAN(**hyperparameters)
tgan.fit(X)
I encountered the following error during fitting :
ValueError: Error when checking input: expected input_6 to have shape (100, 1) but got array with shape (100, 25)
Seems to the input shape at the critic_x
I noticed on the Colab version of Tulog.ipynb , it seems to use tgan.fit(X,X) instead of tgan.fit(X) from github
Thanks
Richard
Paste the command(s) you ran and the output.
If there was a crash, please include the traceback here.
ValueError Traceback (most recent call last)
in
12
13 tgan = TadGAN(**hyperparameters)
---> 14 tgan.fit(X,X)
/app/orion/primitives/tadgan.py in fit(self, X, y, **kwargs)
292 self._build_tadgan(**kwargs)
293
--> 294 self._fit(X, y)
295 self._fitted = True
296
/app/orion/primitives/tadgan.py in _fit(self, X, target)
262 z = np.random.normal(size=(self.batch_size, self.latent_dim, 1))
263 epoch_cx_loss.append(
--> 264 self.critic_x_model.train_on_batch([y, z], [valid, fake, delta]))
265 epoch_cz_loss.append(
266 self.critic_z_model.train_on_batch([x, z], [valid, fake, delta]))
/usr/local/lib/python3.6/dist-packages/keras/engine/training.py in train_on_batch(self, x, y, sample_weight, class_weight, reset_metrics)
1506 x, y,
1507 sample_weight=sample_weight,
-> 1508 class_weight=class_weight)
1509 if self._uses_dynamic_learning_phase():
1510 ins = x + y + sample_weights + [1]
/usr/local/lib/python3.6/dist-packages/keras/engine/training.py in _standardize_user_data(self, x, y, sample_weight, class_weight, check_array_lengths, batch_size)
577 feed_input_shapes,
578 check_batch_axis=False, # Don't enforce the batch size.
--> 579 exception_prefix='input')
580
581 if y is not None:
/usr/local/lib/python3.6/dist-packages/keras/engine/training_utils.py in standardize_input_data(data, names, shapes, check_batch_axis, exception_prefix)
143 ': expected ' + names[i] + ' to have shape ' +
144 str(shape) + ' but got array with shape ' +
--> 145 str(data_shape))
146 return data
147
ValueError: Error when checking input: expected input_6 to have shape (100, 1) but got array with shape (100, 25)
Hi @richardcai88! Thanks for using Orion!
There is one extra step for multivariate TadGAN if you are running the tulog.ipynb, which is that you need to create a y variable that refers to the channel you want to reconstruct.
channel = 0 # index of channel to reconstruct
y = X[:, :, channel:channel+1] # slice it out of X
tgan.fit(X, y)
Let me know if you have any further questions!
Hi, I am facing similar problem with a multivariate dataset of window size 100 and dimensions 25.
Here's my code:
channel = 0
Y = X[:, :, channel:channel+1]
## X.shape - (1225, 100, 25)
## Y.shape - (1225, 100, 1)
from model import hyperparameters
from orion.primitives.tadgan import TadGAN
hyperparameters["epochs"] = 1
hyperparameters["input_shape"] = (100, 1) # based on the window size
hyperparameters["optimizer"] = "keras.optimizers.Adam"
hyperparameters["learning_rate"] = 0.0005
hyperparameters["latent_dim"] = 20
hyperparameters["batch_size"] = 64
tgan = TadGAN(**hyperparameters)
tgan.fit(X, Y)
I get the following error:
ValueError: Exception encountered when calling layer 'critic_x' (type Sequential).
Input 0 of layer "conv1d_24" is incompatible with the layer: expected axis -1 of input shape to have value 25, but received input with shape (None, 100, 1)
Call arguments received by layer 'critic_x' (type Sequential):
• inputs=tf.Tensor(shape=(None, 100, 1), dtype=float64)
• training=None
• mask=None```
Any suggestions?
| gharchive/issue | 2022-09-01T05:41:43 | 2025-04-01T06:45:48.188283 | {
"authors": [
"prithuls",
"richardcai88",
"sarahmish"
],
"repo": "sintel-dev/Orion",
"url": "https://github.com/sintel-dev/Orion/issues/318",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
197393197 | DEF-TEST eats CPU and conses like crazy
I'm on SBCL 1.3.11 on Linux.
(ql:quickload :fiveam)
(use-package :fiveam)
(def-test foo-test ()
#.(cons 'progn (make-list 100 :initial-element '(is (equalp (cons 2 4) (cons 2 4))))))
This causes my SBCL to eat up all of its CPU, then eat up all of its memory, and finally crash and burn because of heap exhaustion.
I #'TIMEd the runs and here are the results. Pay attention to the #'MAKE-LIST list length as it changes from 2 to 64.
CL-USER> (ql:quickload :fiveam)
CL-USER> (use-package :fiveam)
CL-USER> (time (def-test foo-test () #.(cons 'progn (make-list 2 :initial-element '(is (equalp (cons 2 4) (cons 2 4)))))))
Evaluation took:
0.091 seconds of real time
0.088000 seconds of total run time (0.068000 user, 0.020000 system)
[ Run times consist of 0.032 seconds GC time, and 0.056 seconds non-GC time. ]
96.70% CPU
1 form interpreted
113 lambdas converted
228,201,148 processor cycles
10,766,368 bytes consed
FOO-TEST
CL-USER> (time (def-test foo-test () #.(cons 'progn (make-list 4 :initial-element '(is (equalp (cons 2 4) (cons 2 4)))))))
Evaluation took:
0.143 seconds of real time
0.144000 seconds of total run time (0.144000 user, 0.000000 system)
100.70% CPU
1 form interpreted
223 lambdas converted
358,371,848 processor cycles
27,275,104 bytes consed
FOO-TEST
CL-USER> (time (def-test foo-test () #.(cons 'progn (make-list 8 :initial-element '(is (equalp (cons 2 4) (cons 2 4)))))))
Evaluation took:
0.531 seconds of real time
0.480000 seconds of total run time (0.464000 user, 0.016000 system)
[ Run times consist of 0.116 seconds GC time, and 0.364 seconds non-GC time. ]
90.40% CPU
1 form interpreted
443 lambdas converted
1,333,272,895 processor cycles
78,526,736 bytes consed
FOO-TEST
CL-USER> (time (def-test foo-test () #.(cons 'progn (make-list 16 :initial-element '(is (equalp (cons 2 4) (cons 2 4)))))))
Evaluation took:
2.100 seconds of real time
1.980000 seconds of total run time (1.880000 user, 0.100000 system)
[ Run times consist of 0.356 seconds GC time, and 1.624 seconds non-GC time. ]
94.29% CPU
1 form interpreted
883 lambdas converted
5,272,793,292 processor cycles
254,015,344 bytes consed
FOO-TEST
CL-USER> (time (def-test foo-test () #.(cons 'progn (make-list 32 :initial-element '(is (equalp (cons 2 4) (cons 2 4)))))))
Evaluation took:
11.481 seconds of real time
11.316000 seconds of total run time (10.888000 user, 0.428000 system)
[ Run times consist of 1.844 seconds GC time, and 9.472 seconds non-GC time. ]
98.56% CPU
1 form interpreted
1,763 lambdas converted
28,827,998,226 processor cycles
895,486,432 bytes consed
FOO-TEST
CL-USER> (time (def-test foo-test () #.(cons 'progn (make-list 64 :initial-element '(is (equalp (cons 2 4) (cons 2 4)))))))
;; heap exhaustion
Also posted by PuercoPop on #lisp:
(sb-sprof:with-profiling (:mode :alloc :report :flat :max-samples 1000) (def-test foo-test () #.(cons 'progn (make-list 10
:initial-element '(is (equalp (cons 2 4) (cons 2 4)))))))
Profiler sample vector full (96 traces / 10000 samples), doubling the size
Profiler sample vector full (278 traces / 20000 samples), doubling the size
Profiler sample vector full (660 traces / 40000 samples), doubling the size
Number of samples: 1000
Alloc interval: 4 regions (approximately 128 kB)
Total sampling amount: 4000 regions (approximately 128000 kB)
Number of cycles: 0
Sampled threads:
#<SB-THREAD:THREAD "sly-channel-1-mrepl-remote-1" RUNNING {1002451083}>
Self Total Cumul
Nr Count % Count % Count % Calls Function
------------------------------------------------------------------------
1 381 38.1 381 38.1 381 38.1 - SB-C::MAKE-GLOBAL-CONFLICTS
2 289 28.9 298 29.8 670 67.0 - SB-C::COPY-CONSET
3 57 5.7 57 5.7 727 72.7 - SB-C::MAKE-CONSET
4 27 2.7 27 2.7 754 75.4 - SB-REGALLOC::INIT-SB-VECTORS
5 13 1.3 13 1.3 767 76.7 - SB-C::MAKE-REF
6 12 1.2 12 1.2 779 77.9 - MAKE-HASH-TABLE
7 10 1.0 10 1.0 789 78.9 - SB-C::MAKE-TN
8 10 1.0 10 1.0 799 79.9 - SB-C:MAKE-TN-REF
9 9 0.9 9 0.9 808 80.8 - SB-PRETTY::MAKE-PRETTY-STREAM
10 9 0.9 9 0.9 817 81.7 - SB-C::MAKE-LVAR
11 9 0.9 9 0.9 826 82.6 - COPY-STRUCTURE
12 9 0.9 9 0.9 835 83.5 - SB-C::MAKE-IR2-BLOCK
13 7 0.7 7 0.7 842 84.2 - SB-KERNEL:MAKE-SINGLE-VALUE-TYPE
14 7 0.7 7 0.7 849 84.9 - SB-INT:COPY-SSET
15 7 0.7 7 0.7 856 85.6 - SB-C::MAKE-VOP
16 6 0.6 15 1.5 862 86.2 - SB-C::CONSTRAIN-REF-TYPE
17 6 0.6 6 0.6 868 86.8 - SB-C::MAKE-BLOCK
18 6 0.6 6 0.6 874 87.4 - SB-IMPL::ALLOC-HASH-CACHE-LINE/4
19 5 0.5 5 0.5 879 87.9 - SB-C::MAKE-CTRAN
20 5 0.5 5 0.5 884 88.4 - (FLET #:CLEANUP-FUN-450 :IN SB-THREAD::CALL-WITH-RECURSIVE-SYSTEM-LOCK/WITHOUT-GCING)
21 5 0.5 5 0.5 889 88.9 - SB-C::SSET-GROW
22 4 0.4 4 0.4 893 89.3 - SB-IMPL::ALLOCATE-VECTOR-WITH-WIDETAG
23 4 0.4 4 0.4 897 89.7 - SB-C::MAKE-BLOCK-KEY
24 4 0.4 4 0.4 901 90.1 - SB-C::MAKE-LEXENV
25 4 0.4 4 0.4 905 90.5 - SB-C::MAKE-LAMBDA
26 4 0.4 4 0.4 909 90.9 - SB-C::INHERIT-CONSTRAINTS
27 3 0.3 3 0.3 912 91.2 - SB-KERNEL:LIST-SUBSEQ*
28 3 0.3 3 0.3 915 91.5 - SB-C::MAKE-RETURN
29 3 0.3 3 0.3 918 91.8 - REMOVE
30 3 0.3 3 0.3 921 92.1 - SB-C::%CONSET-GROW
31 3 0.3 3 0.3 924 92.4 - SB-KERNEL:TYPE-DIFFERENCE
32 2 0.2 5 0.5 926 92.6 - SB-C::PROPAGATE-LET-ARGS
33 2 0.2 4 0.4 928 92.8 - SB-C::LINK-BLOCKS
34 2 0.2 4 0.4 930 93.0 - SB-C::UPDATE-UVL-LIVE-SETS
35 2 0.2 2 0.2 932 93.2 - SB-C::%LINK-BLOCKS
36 2 0.2 2 0.2 934 93.4 - SB-IMPL::%MAKE-STRING-OUTPUT-STREAM
37 2 0.2 2 0.2 936 93.6 - SB-C::EXTRACT-LET-VARS
38 2 0.2 2 0.2 938 93.8 - SB-PRETTY::START-LOGICAL-BLOCK
39 2 0.2 2 0.2 940 94.0 - SB-C::%MAKE-CAST
40 2 0.2 2 0.2 942 94.2 - SB-INT:PARSE-LAMBDA-LIST
41 2 0.2 2 0.2 944 94.4 - SB-C::MAKE-BIND
42 2 0.2 2 0.2 946 94.6 - SB-C::%MAKE-INTERVAL
43 2 0.2 2 0.2 948 94.8 - SB-C::MAKE-COMBINATION
44 2 0.2 2 0.2 950 95.0 - APPEND
45 2 0.2 2 0.2 952 95.2 - SB-C::PREPARE-ARG-FOR-DERIVE-TYPE
46 2 0.2 2 0.2 954 95.4 - SB-IMPL::ALLOC-HASH-CACHE-LINE/3
47 2 0.2 2 0.2 956 95.6 - SB-KERNEL::%MAKE-VALUES-TYPE
48 2 0.2 2 0.2 958 95.8 - SB-INT:ALLOC-XSET
49 2 0.2 2 0.2 960 96.0 - SB-C::MAP-BLOCK-NLXES
50 2 0.2 2 0.2 962 96.2 - SB-INT:MAKE-SSET
51 1 0.1 233 23.3 963 96.3 - SB-C::FIND-BLOCK-TYPE-CONSTRAINTS
52 1 0.1 9 0.9 964 96.4 - SB-C::REFERENCE-LEAF
53 1 0.1 6 0.6 965 96.5 - SB-C::ADD-EQL-VAR-VAR-CONSTRAINT
54 1 0.1 4 0.4 966 96.6 - SB-KERNEL:%MAKE-ARRAY
55 1 0.1 2 0.2 967 96.7 - (MACRO-FUNCTION DOTIMES)
56 1 0.1 2 0.2 968 96.8 - SB-C::FIND-DFO-AUX
57 1 0.1 1 0.1 969 96.9 - SB-PRETTY::ENQUEUE-NEWLINE
58 1 0.1 1 0.1 970 97.0 - SB-IMPL::STRING-OUTPUT-STREAM-NEW-BUFFER
59 1 0.1 1 0.1 971 97.1 - SB-UNICODE::DECOMPOSE-STRING
60 1 0.1 1 0.1 972 97.2 - (MACRO-FUNCTION POP)
61 1 0.1 1 0.1 973 97.3 - SB-PRETTY::MAKE-BLOCK-END
62 1 0.1 1 0.1 974 97.4 - (LABELS SB-C::WALK :IN SB-C::DEBUG-NAME)
63 1 0.1 1 0.1 975 97.5 - (LABELS SB-C::HANDLE-FUNCTIONAL :IN SB-C::CALL-WITH-BLOCK-EXTERNAL-FUNCTIONALS)
64 1 0.1 1 0.1 976 97.6 - SB-FORMAT::EXPAND-DIRECTIVE
65 1 0.1 1 0.1 977 97.7 - SB-C::MAKE-IF
66 1 0.1 1 0.1 978 97.8 - SB-C::MAKE-TAIL-SET
67 1 0.1 1 0.1 979 97.9 - SB-C::MAKE-ENTRY
68 1 0.1 1 0.1 980 98.0 - SB-KERNEL::UNION-UNPARSE-TYPE-METHOD
69 1 0.1 1 0.1 981 98.1 - SB-KERNEL:%MAKE-SYMBOL
70 1 0.1 1 0.1 982 98.2 - SB-KERNEL::NUMBER-UNPARSE-TYPE-METHOD
71 1 0.1 1 0.1 983 98.3 - (MACRO-FUNCTION COND)
72 1 0.1 1 0.1 984 98.4 - SB-C::MAKE-EXIT
73 1 0.1 1 0.1 985 98.5 - (FLET SB-IMPL::FROB-DO-BODY :IN "SYS:SRC;CODE;DEFBOOT.LISP")
74 1 0.1 1 0.1 986 98.6 - SB-C::ENSURE-SOURCE-PATH
75 1 0.1 1 0.1 987 98.7 - SB-KERNEL::UNION-COMPLEX-SUBTYPEP-ARG2
76 1 0.1 1 0.1 988 98.8 - SB-C::ADD-LVAR-USE
77 1 0.1 1 0.1 989 98.9 - SB-KERNEL::FIXED-VALUES-OP
78 1 0.1 1 0.1 990 99.0 - SB-C::MAKE-CONSTANT
79 1 0.1 1 0.1 991 99.1 - (MACRO-FUNCTION PROG1)
80 1 0.1 1 0.1 992 99.2 - SB-C::MAYBE-NEGATE-CHECK
81 1 0.1 1 0.1 993 99.3 - SB-C::PRECISE-ADD-COMPLEMENT-CONSTRAINTS
82 1 0.1 1 0.1 994 99.4 - (FLET SB-C::BODY-FUN :IN SB-C::ADD-EQL-VAR-VAR-CONSTRAINT)
83 1 0.1 1 0.1 995 99.5 - SB-C::NCONC-NEW
84 1 0.1 1 0.1 996 99.6 - SB-C::EMIT-CLEANUPS
85 1 0.1 1 0.1 997 99.7 - "foreign function pthread_sigmask"
86 1 0.1 1 0.1 998 99.8 - SB-C::MAKE-IR2-LVAR
87 1 0.1 1 0.1 999 99.9 - SB-KERNEL:%ADJOIN-EQ
88 1 0.1 1 0.1 1000 100.0 - SB-REGALLOC::GROW-SC
89 0 0.0 1000 100.0 1000 100.0 - SB-C::%COMPILE
90 0 0.0 1000 100.0 1000 100.0 - (FLET #:WITHOUT-INTERRUPTS-BODY-387 :IN SB-THREAD::CALL-WITH-RECURSIVE-LOCK)
91 0 0.0 1000 100.0 1000 100.0 - SB-THREAD::CALL-WITH-RECURSIVE-LOCK
92 0 0.0 1000 100.0 1000 100.0 - (LAMBDA NIL :IN SB-C::ACTUALLY-COMPILE)
93 0 0.0 1000 100.0 1000 100.0 - (FLET SB-C::WITH-IT :IN SB-C::%WITH-COMPILATION-UNIT)
94 0 0.0 1000 100.0 1000 100.0 - SB-C::ACTUALLY-COMPILE
95 0 0.0 1000 100.0 1000 100.0 - SB-C:COMPILE-IN-LEXENV
96 0 0.0 1000 100.0 1000 100.0 - SB-IMPL::%SIMPLE-EVAL
97 0 0.0 1000 100.0 1000 100.0 - SB-INT:SIMPLE-EVAL-IN-LEXENV
98 0 0.0 1000 100.0 1000 100.0 - EVAL
99 0 0.0 1000 100.0 1000 100.0 - IT.BESE.FIVEAM::REGISTER-TEST
100 0 0.0 1000 100.0 1000 100.0 - "Unknown component: #x22A12050"
101 0 0.0 1000 100.0 1000 100.0 - (LAMBDA NIL :IN SLYNK-MREPL::MREPL-EVAL-1)
102 0 0.0 1000 100.0 1000 100.0 - SLYNK::CALL-WITH-RETRY-RESTART
103 0 0.0 1000 100.0 1000 100.0 - (LAMBDA NIL :IN SLYNK::CALL-WITH-LISTENER)
104 0 0.0 1000 100.0 1000 100.0 - SLYNK::CALL-WITH-BINDINGS
105 0 0.0 1000 100.0 1000 100.0 - SLYNK::CALL-WITH-LISTENER
106 0 0.0 1000 100.0 1000 100.0 - SLYNK-MREPL::MREPL-EVAL-1
107 0 0.0 1000 100.0 1000 100.0 - SLYNK-MREPL::MREPL-EVAL
108 0 0.0 1000 100.0 1000 100.0 - (SB-PCL::FAST-METHOD SLYNK::CHANNEL-SEND (SLYNK-MREPL::MREPL (EQL (QUOTE :PROCESS)) T))
109 0 0.0 1000 100.0 1000 100.0 - (LAMBDA (SB-PCL::.PV. SB-PCL::.NEXT-METHOD-CALL. SB-PCL::.ARG0. SB-PCL::.ARG1. SB-PCL::.ARG2.))
110 0 0.0 1000 100.0 1000 100.0 - SLYNK::PROCESS-REQUESTS
111 0 0.0 1000 100.0 1000 100.0 - (LAMBDA NIL :IN SLYNK::SPAWN-CHANNEL-THREAD)
112 0 0.0 1000 100.0 1000 100.0 - SLYNK-SBCL::CALL-WITH-BREAK-HOOK
113 0 0.0 1000 100.0 1000 100.0 - (FLET SLYNK-BACKEND:CALL-WITH-DEBUGGER-HOOK :IN "/home/puercopop/.emacs.d/site-lisp/sly/slynk/backend/sbcl.lisp")
114 0 0.0 1000 100.0 1000 100.0 - SLYNK-BACKEND:CALL-WITH-DEBUGGER-HOOK
115 0 0.0 1000 100.0 1000 100.0 - (FLET #:FORM-FUN-3522 :IN SLYNK::SPAWN-CHANNEL-THREAD)
116 0 0.0 1000 100.0 1000 100.0 - (FLET #:WITHOUT-INTERRUPTS-BODY-1158 :IN SB-THREAD::INITIAL-THREAD-FUNCTION-TRAMPOLINE)
117 0 0.0 1000 100.0 1000 100.0 - (FLET SB-THREAD::WITH-MUTEX-THUNK :IN SB-THREAD::INITIAL-THREAD-FUNCTION-TRAMPOLINE)
118 0 0.0 1000 100.0 1000 100.0 - (FLET #:WITHOUT-INTERRUPTS-BODY-359 :IN SB-THREAD::CALL-WITH-MUTEX)
119 0 0.0 1000 100.0 1000 100.0 - SB-THREAD::CALL-WITH-MUTEX
120 0 0.0 1000 100.0 1000 100.0 - SB-THREAD::INITIAL-THREAD-FUNCTION-TRAMPOLINE
121 0 0.0 1000 100.0 1000 100.0 - "foreign function call_into_lisp"
122 0 0.0 1000 100.0 1000 100.0 - "foreign function new_thread_trampoline"
123 0 0.0 970 97.0 1000 100.0 - SB-C::COMPILE-COMPONENT
124 0 0.0 499 49.9 1000 100.0 - SB-C::IR1-PHASES
125 0 0.0 460 46.0 1000 100.0 - SB-C::%COMPILE-COMPONENT
126 0 0.0 394 39.4 1000 100.0 - SB-C::CONSTRAINT-PROPAGATE
127 0 0.0 381 38.1 1000 100.0 - SB-C::ADD-GLOBAL-CONFLICT
128 0 0.0 381 38.1 1000 100.0 - SB-C::LIFETIME-ANALYZE
129 0 0.0 380 38.0 1000 100.0 - SB-C::SETUP-ENVIRONMENT-TN-CONFLICT
130 0 0.0 380 38.0 1000 100.0 - SB-C::SETUP-ENVIRONMENT-TN-CONFLICTS
131 0 0.0 378 37.8 1000 100.0 - SB-C::FIND-AND-PROPAGATE-CONSTRAINTS
132 0 0.0 338 33.8 1000 100.0 - SB-C::SETUP-ENVIRONMENT-LIVE-CONFLICTS
133 0 0.0 144 14.4 1000 100.0 - SB-C::COMPUTE-BLOCK-IN
134 0 0.0 114 11.4 1000 100.0 - SB-C::UPDATE-BLOCK-IN
135 0 0.0 96 9.6 1000 100.0 - SB-C::IR1-OPTIMIZE
136 0 0.0 96 9.6 1000 100.0 - SB-C::IR1-OPTIMIZE-UNTIL-DONE
137 0 0.0 94 9.4 1000 100.0 - SB-C::IR1-OPTIMIZE-BLOCK
138 0 0.0 90 9.0 1000 100.0 - SB-C::IR1-CONVERT-LAMBDA-BODY
139 0 0.0 87 8.7 1000 100.0 - SB-C::IR1-CONVERT
140 0 0.0 87 8.7 1000 100.0 - SB-C::IR1-CONVERT-LAMBDA
141 0 0.0 85 8.5 1000 100.0 - SB-C::IR1-CONVERT-AUX-BINDINGS
142 0 0.0 85 8.5 1000 100.0 - SB-C::IR1-CONVERT-SPECIAL-BINDINGS
143 0 0.0 85 8.5 1000 100.0 - SB-C::IR1-CONVERT-PROGN-BODY
144 0 0.0 85 8.5 1000 100.0 - SB-C::IR1-OPTIMIZE-COMBINATION
145 0 0.0 79 7.9 1000 100.0 - SB-C::IR1-CONVERT-LET
146 0 0.0 77 7.7 1000 100.0 - SB-C::%PROCESSING-DECLS
147 0 0.0 77 7.7 1000 100.0 - SB-C::IR1-CONVERT-GLOBAL-FUNCTOID
148 0 0.0 76 7.6 1000 100.0 - SB-C::IR1-TRANSFORM
149 0 0.0 75 7.5 1000 100.0 - SB-C::%WITH-IR1-ENVIRONMENT-FROM-NODE
150 0 0.0 73 7.3 1000 100.0 - (LAMBDA (SB-C::NEXT SB-C::RESULT SB-C::POST-BINDING-LEXENV) :IN SB-C::IR1-CONVERT-LET)
151 0 0.0 67 6.7 1000 100.0 - SB-C::TRANSFORM-CALL
152 0 0.0 66 6.6 1000 100.0 - (FLET SB-C::CLOSURE-NEEDING-IR1-ENVIRONMENT-FROM-NODE :IN SB-C::TRANSFORM-CALL)
153 0 0.0 65 6.5 1000 100.0 - SB-C::IR1-CONVERT-COMBINATION-CHECKING-TYPE
154 0 0.0 65 6.5 1000 100.0 - SB-C::IR1-CONVERT-BLOCK
155 0 0.0 64 6.4 1000 100.0 - SB-C::IR1-CONVERT-COMBINATION
156 0 0.0 58 5.8 1000 100.0 - SB-C::ADD-TEST-CONSTRAINTS
157 0 0.0 57 5.7 1000 100.0 - SB-C::IR1-CONVERT-COMBINATION-ARGS
158 0 0.0 56 5.6 1000 100.0 - SB-C::IR1-CONVERT-IF
159 0 0.0 56 5.6 1000 100.0 - SB-C::IR1-CONVERT-INLINE-LAMBDA
160 0 0.0 52 5.2 1000 100.0 - SB-C::IR1-CONVERT-PROGN
161 0 0.0 43 4.3 1000 100.0 - SB-C::IR1-CONVERT-MORE
162 0 0.0 43 4.3 1000 100.0 - SB-C::IR1-CONVERT-HAIRY-LAMBDA
163 0 0.0 42 4.2 1000 100.0 - SB-C::CONVERT-TO-ENVIRONMENT-TN
164 0 0.0 42 4.2 1000 100.0 - SB-C::CONFLICTIZE-SAVE-P-VOP
165 0 0.0 42 4.2 1000 100.0 - SB-C::CONFLICT-ANALYZE-1-BLOCK
166 0 0.0 42 4.2 1000 100.0 - SB-C::LIFETIME-POST-PASS
167 0 0.0 38 3.8 1000 100.0 - SB-C::IR1-CONVERT-TAGBODY
168 0 0.0 37 3.7 1000 100.0 - SB-C::IR1-CONVERT-FUNCTION
169 0 0.0 35 3.5 1000 100.0 - SB-C::FUN-NAME-LEAF
170 0 0.0 33 3.3 1000 100.0 - SB-C::CONSTRAINT-PROPAGATE-IN-BLOCK
171 0 0.0 29 2.9 1000 100.0 - SB-C::IR1-CONVERT-MULTIPLE-VALUE-PROG1
172 0 0.0 29 2.9 1000 100.0 - SB-C::IR1-CONVERT-LABELS
173 0 0.0 29 2.9 1000 100.0 - SB-C::MAKE-FUNCTIONAL-FROM-TOPLEVEL-LAMBDA
174 0 0.0 28 2.8 1000 100.0 - SB-REGALLOC:PACK
175 0 0.0 22 2.2 1000 100.0 - SB-C::IR2-CONVERT-BLOCK
176 0 0.0 22 2.2 1000 100.0 - SB-C:IR2-CONVERT
177 0 0.0 20 2.0 1000 100.0 - SB-INT:STRINGIFY-OBJECT
178 0 0.0 20 2.0 1000 100.0 - PRIN1-TO-STRING
179 0 0.0 20 2.0 1000 100.0 - SB-C::EMIT-VOP
180 0 0.0 20 2.0 1000 100.0 - SB-C::EMIT-AND-INSERT-VOP
181 0 0.0 19 1.9 1000 100.0 - SB-C::IR1-CONVERT-RETURN-FROM
182 0 0.0 18 1.8 1000 100.0 - (LABELS SB-IMPL::CHECK-IT :IN SB-KERNEL:OUTPUT-OBJECT)
183 0 0.0 15 1.5 1000 100.0 - (LABELS SB-IMPL::PRINT-IT :IN SB-KERNEL:OUTPUT-OBJECT)
184 0 0.0 15 1.5 1000 100.0 - (LABELS SB-IMPL::HANDLE-IT :IN SB-KERNEL:OUTPUT-OBJECT)
185 0 0.0 14 1.4 1000 100.0 - SB-C::IR1-CONVERT-THE
186 0 0.0 13 1.3 1000 100.0 - SB-C::CONVERT-MORE-ENTRY
187 0 0.0 13 1.3 1000 100.0 - SB-C::LTN-ANALYZE
188 0 0.0 12 1.2 1000 100.0 - SB-C::IR1-CONVERT-FLET
189 0 0.0 12 1.2 1000 100.0 - SB-C::IR1-CONVERT-UNWIND-PROTECT
190 0 0.0 12 1.2 1000 100.0 - SB-C::FIND-OR-CREATE-CONSTRAINT
191 0 0.0 11 1.1 1000 100.0 - SB-C::THE-IN-POLICY
192 0 0.0 11 1.1 1000 100.0 - SB-C::IR1-CONVERT-MULTIPLE-VALUE-CALL
193 0 0.0 11 1.1 1000 100.0 - SB-C::LOCALL-ANALYZE-COMPONENT
194 0 0.0 11 1.1 1000 100.0 - SB-C::REGISTER-CONSTRAINT
195 0 0.0 10 1.0 1000 100.0 - SB-C::REFERENCE-TN
196 0 0.0 9 0.9 1000 100.0 - SB-C::DERIVE-NODE-TYPE
197 0 0.0 9 0.9 1000 100.0 - SB-C::GENERATE-TYPE-CHECKS
198 0 0.0 9 0.9 1000 100.0 - SB-C::IR2-CONVERT-FIXED-FULL-CALL
199 0 0.0 9 0.9 1000 100.0 - SB-C::IR2-CONVERT-FULL-CALL
200 0 0.0 9 0.9 1000 100.0 - SB-C::EMIT-MOVE
201 0 0.0 8 0.8 1000 100.0 - SB-C::LET-CONVERT
202 0 0.0 8 0.8 1000 100.0 - SB-C::MAYBE-LET-CONVERT
203 0 0.0 8 0.8 1000 100.0 - SB-C::MAYBE-ADD-EQL-VAR-LVAR-CONSTRAINT
204 0 0.0 8 0.8 1000 100.0 - (FLET SB-C::CLOSURE-NEEDING-IR1-ENVIRONMENT-FROM-NODE :IN SB-C::FILTER-LVAR)
205 0 0.0 8 0.8 1000 100.0 - SB-C::FILTER-LVAR
206 0 0.0 8 0.8 1000 100.0 - SB-C::CONVERT-TYPE-CHECK
207 0 0.0 8 0.8 1000 100.0 - SB-C::FIND-DOMINATORS
208 0 0.0 7 0.7 1000 100.0 - SB-C::CAREFUL-EXPAND-MACRO
209 0 0.0 7 0.7 1000 100.0 - SB-C::NODE-ENDS-BLOCK
210 0 0.0 7 0.7 1000 100.0 - SB-C::IR1-CONVERT-VAR
211 0 0.0 7 0.7 1000 100.0 - SB-C::IR1-OPTIMIZE-SET
212 0 0.0 7 0.7 1000 100.0 - SB-C::IR1-CONVERT-LOCALLY
213 0 0.0 6 0.6 1000 100.0 - SB-C::REFERENCE-CONSTANT
214 0 0.0 6 0.6 1000 100.0 - SB-C::GENERATE-OPTIONAL-DEFAULT-ENTRY
215 0 0.0 6 0.6 1000 100.0 - SB-C::IR1-CONVERT-HAIRY-ARGS
216 0 0.0 6 0.6 1000 100.0 - SB-C::CTRAN-STARTS-BLOCK
217 0 0.0 6 0.6 1000 100.0 - SB-KERNEL:VALUES-TYPE-INTERSECTION
218 0 0.0 5 0.5 1000 100.0 - SB-C::IR1-CONVERT-FBINDINGS
219 0 0.0 5 0.5 1000 100.0 - (LAMBDA (SB-C::NEXT SB-C::RESULT #:LEXENV828) :IN SB-C::IR1-CONVERT-FLET)
220 0 0.0 5 0.5 1000 100.0 - (LABELS #:BODY-NAME-2628 :IN SB-PRETTY::CALL-LOGICAL-BLOCK-PRINTER)
221 0 0.0 5 0.5 1000 100.0 - (FLET #:WITH-PRETTY-STREAM2626 :IN SB-PRETTY::CALL-LOGICAL-BLOCK-PRINTER)
222 0 0.0 5 0.5 1000 100.0 - SB-PRETTY::CALL-LOGICAL-BLOCK-PRINTER
223 0 0.0 5 0.5 1000 100.0 - SB-PRETTY::PPRINT-FUN-CALL
224 0 0.0 5 0.5 1000 100.0 - (FLET SB-C::CLOSURE-NEEDING-IR1-ENVIRONMENT-FROM-NODE :IN SB-C::MAKE-XEP)
225 0 0.0 5 0.5 1000 100.0 - SB-C::REFERENCE-ENTRY-POINT
226 0 0.0 5 0.5 1000 100.0 - SB-C::LOCALL-ANALYZE-FUN-1
227 0 0.0 5 0.5 1000 100.0 - SB-THREAD::CALL-WITH-RECURSIVE-SYSTEM-LOCK/WITHOUT-GCING
228 0 0.0 5 0.5 1000 100.0 - (FLET #:BODY-FUN-694 :IN SB-KERNEL:%PUTHASH)
229 0 0.0 5 0.5 1000 100.0 - SB-KERNEL:%PUTHASH
230 0 0.0 5 0.5 1000 100.0 - SB-INT:SSET-ADJOIN
231 0 0.0 5 0.5 1000 100.0 - SB-C::IR2-CONVERT-LOCAL-CALL
232 0 0.0 5 0.5 1000 100.0 - SB-C::IR2-OPTIMIZE
233 0 0.0 4 0.4 1000 100.0 - SB-C::IR1-CONVERT-TRULY-THE
234 0 0.0 4 0.4 1000 100.0 - SB-C::IR1-CONVERT-LET*
235 0 0.0 4 0.4 1000 100.0 - (FLET #:PPRINT-BLOCK :IN SB-PRETTY::PPRINT-FUN-CALL)
236 0 0.0 4 0.4 1000 100.0 - SB-C::IR1-CONVERT-%WITHIN-CLEANUP
237 0 0.0 4 0.4 1000 100.0 - SB-C::SETQ-VAR
238 0 0.0 4 0.4 1000 100.0 - SB-C::IR1-CONVERT-SETQ
239 0 0.0 4 0.4 1000 100.0 - SB-C::INSERT-LET-BODY
240 0 0.0 4 0.4 1000 100.0 - SB-C::IR1-CONVERT-%FUNCALL
241 0 0.0 4 0.4 1000 100.0 - SB-KERNEL::ARGS-TYPE-OP
242 0 0.0 4 0.4 1000 100.0 - SB-KERNEL::VALUES-TYPE-OP
243 0 0.0 4 0.4 1000 100.0 - SB-C::STACK-ANALYZE
244 0 0.0 4 0.4 1000 100.0 - SB-C::IR2-CONVERT-REF
245 0 0.0 4 0.4 1000 100.0 - SB-C:MAKE-RESTRICTED-TN
246 0 0.0 4 0.4 1000 100.0 - (LABELS SB-C::LINK-2BLOCKS :IN SB-C::INITIALIZE-IR2-BLOCKS-FLOW-INFO)
247 0 0.0 4 0.4 1000 100.0 - SB-C::INITIALIZE-IR2-BLOCKS-FLOW-INFO
248 0 0.0 3 0.3 1000 100.0 - SB-C::ASSERT-LVAR-TYPE
249 0 0.0 3 0.3 1000 100.0 - (LAMBDA (SB-C::ARG TYPE) :IN SB-C::ASSERT-CALL-TYPE)
250 0 0.0 3 0.3 1000 100.0 - SB-C::MAP-COMBINATION-ARGS-AND-TYPES
251 0 0.0 3 0.3 1000 100.0 - SB-C::ASSERT-CALL-TYPE
252 0 0.0 3 0.3 1000 100.0 - SB-C::VALIDATE-CALL-TYPE
253 0 0.0 3 0.3 1000 100.0 - SB-C::STRIP-LAMBDA-LIST
254 0 0.0 3 0.3 1000 100.0 - SB-C::CHECK-ARG-TYPE
255 0 0.0 3 0.3 1000 100.0 - SB-C::CHECK-FIXED-AND-REST
256 0 0.0 3 0.3 1000 100.0 - SB-C::VALID-FUN-USE
257 0 0.0 3 0.3 1000 100.0 - SB-KERNEL:TYPE=
258 0 0.0 3 0.3 1000 100.0 - SB-KERNEL:TYPE/=
259 0 0.0 3 0.3 1000 100.0 - SB-C::TYPE-APPROXIMATE-INTERVAL
260 0 0.0 3 0.3 1000 100.0 - SB-C::MOVE-RETURN-USES
261 0 0.0 3 0.3 1000 100.0 - SB-C::MOVE-RETURN-STUFF
262 0 0.0 3 0.3 1000 100.0 - SB-C::CHANGE-REF-LEAF
263 0 0.0 3 0.3 1000 100.0 - SB-C:MAKE-NORMAL-TN
264 0 0.0 3 0.3 1000 100.0 - SB-C::LTN-ANALYZE-BLOCK
265 0 0.0 3 0.3 1000 100.0 - SB-C:MAKE-WIRED-TN
266 0 0.0 3 0.3 1000 100.0 - SB-C::IR2-CONVERT-LET
267 0 0.0 3 0.3 1000 100.0 - SB-C::IR2-CONVERT-FULL-CALL-ARGS
268 0 0.0 3 0.3 1000 100.0 - SB-C::EMIT-ARG-MOVES
269 0 0.0 3 0.3 1000 100.0 - SB-C::EMIT-MOVES-AND-COERCIONS
270 0 0.0 3 0.3 1000 100.0 - SB-C::SELECT-REPRESENTATIONS
271 0 0.0 2 0.2 1000 100.0 - SB-C::IR1-CONVERT-QUOTE
272 0 0.0 2 0.2 1000 100.0 - (LABELS SB-IMPL::RECURSE :IN SB-IMPL::%MAKE-SYMBOL-NAME)
273 0 0.0 2 0.2 1000 100.0 - GENSYM
274 0 0.0 2 0.2 1000 100.0 - SB-FORMAT::%FORMATTER
275 0 0.0 2 0.2 1000 100.0 - (SB-C:DEFTRANSFORM FORMAT)
276 0 0.0 2 0.2 1000 100.0 - (FLET SB-C::CLOSURE-NEEDING-IR1-ENVIRONMENT-FROM-NODE :IN SB-C::INSERT-CAST-BEFORE)
277 0 0.0 2 0.2 1000 100.0 - SB-C::CONVERT-OPTIONAL-ENTRY
278 0 0.0 2 0.2 1000 100.0 - SB-KERNEL:VALUES-SUBTYPEP
279 0 0.0 2 0.2 1000 100.0 - SB-C::IR1-OPTIMIZE-CAST
280 0 0.0 2 0.2 1000 100.0 - (SB-C:DEFTRANSFORM <)
281 0 0.0 2 0.2 1000 100.0 - SB-C::DELETE-REF
282 0 0.0 2 0.2 1000 100.0 - SB-KERNEL::MAKE-VALUES-TYPE-CACHED
283 0 0.0 2 0.2 1000 100.0 - SB-KERNEL:CSUBTYPEP
284 0 0.0 2 0.2 1000 100.0 - SB-KERNEL::%COERCE-TO-VALUES
285 0 0.0 2 0.2 1000 100.0 - SB-C::CONSET-INTERSECTION
286 0 0.0 2 0.2 1000 100.0 - SB-KERNEL:%MAP-FOR-EFFECT-ARITY-1
287 0 0.0 2 0.2 1000 100.0 - SB-C::FIND-DFO
288 0 0.0 2 0.2 1000 100.0 - SB-C::DFO-AS-NEEDED
289 0 0.0 2 0.2 1000 100.0 - SB-C::LTN-ANALYZE-KNOWN-CALL
290 0 0.0 2 0.2 1000 100.0 - SB-C::IR2-CONVERT-TAIL-LOCAL-CALL
291 0 0.0 2 0.2 1000 100.0 - SB-C::COPY-PROPAGATE
292 0 0.0 1 0.1 1000 100.0 - PPRINT-NEWLINE
293 0 0.0 1 0.1 1000 100.0 - SB-IMPL::STRING-SOUT
294 0 0.0 1 0.1 1000 100.0 - SB-IMPL::%WRITE-STRING
295 0 0.0 1 0.1 1000 100.0 - (MACRO-FUNCTION DOLIST)
296 0 0.0 1 0.1 1000 100.0 - SB-UNICODE:NORMALIZE-STRING
297 0 0.0 1 0.1 1000 100.0 - SB-UNICODE:NORMALIZED-P
298 0 0.0 1 0.1 1000 100.0 - SB-KERNEL:OUTPUT-SYMBOL-NAME
299 0 0.0 1 0.1 1000 100.0 - SB-PRETTY::END-LOGICAL-BLOCK
300 0 0.0 1 0.1 1000 100.0 - SB-C::DEBUG-NAME
301 0 0.0 1 0.1 1000 100.0 - SB-C::LOCALL-ANALYZE-CLAMBDAS-UNTIL-DONE
302 0 0.0 1 0.1 1000 100.0 - SB-C::RECORD-COMPONENT-XREFS
303 0 0.0 1 0.1 1000 100.0 - SB-FORMAT::EXPAND-DIRECTIVE-LIST
304 0 0.0 1 0.1 1000 100.0 - SB-FORMAT::EXPAND-CONTROL-STRING
305 0 0.0 1 0.1 1000 100.0 - SB-C::PROPAGATE-FROM-SETS
306 0 0.0 1 0.1 1000 100.0 - SB-C::+-DERIVE-TYPE-AUX
307 0 0.0 1 0.1 1000 100.0 - SB-C::TWO-ARG-DERIVE-TYPE
308 0 0.0 1 0.1 1000 100.0 - SB-INT:MAKE-LAMBDA-LIST
309 0 0.0 1 0.1 1000 100.0 - SB-C::SSET-UNION
310 0 0.0 1 0.1 1000 100.0 - SB-C::MERGE-LETS
311 0 0.0 1 0.1 1000 100.0 - (SB-C:DEFTRANSFORM >=)
312 0 0.0 1 0.1 1000 100.0 - SB-C::CHANGE-BLOCK-SUCCESSOR
313 0 0.0 1 0.1 1000 100.0 - SB-C::UNLINK-NODE
314 0 0.0 1 0.1 1000 100.0 - SB-C::ENSURE-BLOCK-START
315 0 0.0 1 0.1 1000 100.0 - SB-C::DELETE-BLOCK
316 0 0.0 1 0.1 1000 100.0 - SB-C::CLEAN-COMPONENT
317 0 0.0 1 0.1 1000 100.0 - SB-C::JOIN-BLOCKS
318 0 0.0 1 0.1 1000 100.0 - SB-C::JOIN-SUCCESSOR-IF-POSSIBLE
319 0 0.0 1 0.1 1000 100.0 - SB-C::SUBSTITUTE-SINGLE-USE-LVAR
320 0 0.0 1 0.1 1000 100.0 - (FLET SB-C::DO-1-USE :IN SB-C::MOVE-RETURN-USES)
321 0 0.0 1 0.1 1000 100.0 - (FLET SB-C::FROB :IN SB-C::DELETE-OPTIONAL-DISPATCH)
322 0 0.0 1 0.1 1000 100.0 - SB-C::DELETE-OPTIONAL-DISPATCH
323 0 0.0 1 0.1 1000 100.0 - SB-C::DELETE-LAMBDA
324 0 0.0 1 0.1 1000 100.0 - SB-C::CONVERT-CALL
325 0 0.0 1 0.1 1000 100.0 - SB-C::CONVERT-HAIRY-CALL
326 0 0.0 1 0.1 1000 100.0 - SB-C::CONVERT-CALL-IF-POSSIBLE
327 0 0.0 1 0.1 1000 100.0 - SB-C::PROPAGATE-FUN-CHANGE
328 0 0.0 1 0.1 1000 100.0 - SB-C::PROPAGATE-TO-REFS
329 0 0.0 1 0.1 1000 100.0 - SB-KERNEL::TYPE=-LIST
330 0 0.0 1 0.1 1000 100.0 - SB-KERNEL::TYPE=-ARGS
331 0 0.0 1 0.1 1000 100.0 - SB-C::IR1-CONVERT-%%ALLOCATE-CLOSURES
332 0 0.0 1 0.1 1000 100.0 - (FLET SB-C::DO-1-USE :IN SB-C::SUBSTITUTE-LVAR-USES)
333 0 0.0 1 0.1 1000 100.0 - SB-C::SUBSTITUTE-LVAR-USES
334 0 0.0 1 0.1 1000 100.0 - SB-C::DELETE-FILTER
335 0 0.0 1 0.1 1000 100.0 - SB-C::DELETE-CAST
336 0 0.0 1 0.1 1000 100.0 - SB-IMPL::EXTEND-VECTOR
337 0 0.0 1 0.1 1000 100.0 - VECTOR-PUSH-EXTEND
338 0 0.0 1 0.1 1000 100.0 - SB-IMPL::DATA-VECTOR-FROM-INITS
339 0 0.0 1 0.1 1000 100.0 - SB-C::CONSET-UNION
340 0 0.0 1 0.1 1000 100.0 - SB-C::FIND-CONSTANT
341 0 0.0 1 0.1 1000 100.0 - SB-C::CAST-CHECK-TYPES
342 0 0.0 1 0.1 1000 100.0 - SB-C::SUBSTITUTE-LEAF-IF
343 0 0.0 1 0.1 1000 100.0 - SB-C::ADD-COMBINATION-TEST-CONSTRAINTS
344 0 0.0 1 0.1 1000 100.0 - SB-C::MAKE-CAST
345 0 0.0 1 0.1 1000 100.0 - (FLET SB-C::ENQUEUE :IN SB-C::FIND-AND-PROPAGATE-CONSTRAINTS)
346 0 0.0 1 0.1 1000 100.0 - SB-C::FIND-CLEANUP-POINTS
347 0 0.0 1 0.1 1000 100.0 - SB-C::PHYSENV-ANALYZE
348 0 0.0 1 0.1 1000 100.0 - SB-C::ASSIGN-LAMBDA-VAR-TNS
349 0 0.0 1 0.1 1000 100.0 - SB-C::GTN-ANALYZE
350 0 0.0 1 0.1 1000 100.0 - "foreign function interrupt_handle_pending"
351 0 0.0 1 0.1 1000 100.0 - "foreign function handle_trap"
352 0 0.0 1 0.1 1000 100.0 - SB-C::SET-TAIL-LOCAL-CALL-SUCCESSOR
353 0 0.0 1 0.1 1000 100.0 - SB-C::LTN-ANALYZE-LOCAL-CALL
354 0 0.0 1 0.1 1000 100.0 - SB-C::ANNOTATE-1-VALUE-LVAR
355 0 0.0 1 0.1 1000 100.0 - SB-C::FIND-LOWEST-COMMON-DOMINATOR
356 0 0.0 1 0.1 1000 100.0 - SB-C::BACK-PROPAGATE-ONE-DX-LVAR
357 0 0.0 1 0.1 1000 100.0 - SB-C::BACK-PROPAGATE-DX-LVARS
358 0 0.0 1 0.1 1000 100.0 - SB-C::EMIT-INITS
359 0 0.0 1 0.1 1000 100.0 - SB-C::IR2-CONVERT-FIXED-ALLOCATION
360 0 0.0 1 0.1 1000 100.0 - SB-C::FINISH-IR2-BLOCK
361 0 0.0 1 0.1 1000 100.0 - SB-C::EMIT-NLX-START
362 0 0.0 1 0.1 1000 100.0 - SB-C::MOVE-RESULTS-COERCED
363 0 0.0 1 0.1 1000 100.0 - SB-C::MOVE-LVAR-RESULT
364 0 0.0 1 0.1 1000 100.0 - SB-C::IR2-CONVERT-CONDITIONAL
365 0 0.0 1 0.1 1000 100.0 - SB-C::IR2-CONVERT-TEMPLATE
366 0 0.0 1 0.1 1000 100.0 - SB-C::INIT-COPY-SETS
367 0 0.0 1 0.1 1000 100.0 - SB-C::COPY-FLOW-ANALYSIS
368 0 0.0 1 0.1 1000 100.0 - (LABELS SB-C::MARK-2BLOCK :IN SB-C::DELETE-UNUSED-IR2-BLOCKS)
369 0 0.0 1 0.1 1000 100.0 - SB-C::DELETE-UNUSED-IR2-BLOCKS
370 0 0.0 1 0.1 1000 100.0 - SB-C::EMIT-MOVE-ARG-TEMPLATE
371 0 0.0 1 0.1 1000 100.0 - SB-C::CONVERT-TO-GLOBAL
372 0 0.0 1 0.1 1000 100.0 - SB-C::INIT-GLOBAL-CONFLICT-KIND
373 0 0.0 1 0.1 1000 100.0 - SB-C::LIFETIME-PRE-PASS
374 0 0.0 1 0.1 1000 100.0 - SB-REGALLOC::PACK-WIRED-TN
375 0 0.0 1 0.1 1000 100.0 - SB-REGALLOC::PACK-GREEDY
------------------------------------------------------------------------
0 0.0 elsewhere
process-failure is a macro for some reason, changing it to
(defun process-failure (&rest args)
(with-simple-restart (ignore-failure "Continue the test run.")
(apply #'error 'check-failure args))
(apply #'add-result 'test-failure args))
should do the trick.
Stassats' fix PR #39 seems to have worked. Thanks both of you.
| gharchive/issue | 2016-12-23T16:08:36 | 2025-04-01T06:45:48.201135 | {
"authors": [
"phoe",
"sionescu",
"stassats"
],
"repo": "sionescu/fiveam",
"url": "https://github.com/sionescu/fiveam/issues/38",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1419959565 | Update sbt-scoverage to 2.0.6
Updates org.scoverage:sbt-scoverage from 1.6.0 to 2.0.6.
GitHub Release Notes - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scoverage", artifactId = "sbt-scoverage" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "@monthly" },
dependency = { groupId = "org.scoverage", artifactId = "sbt-scoverage" }
}]
labels: sbt-plugin-update, early-semver-major, semver-spec-major, commit-count:1
Superseded by #50.
| gharchive/pull-request | 2022-10-23T22:28:18 | 2025-04-01T06:45:48.242880 | {
"authors": [
"scala-steward"
],
"repo": "sirgraystar/mandyville-modelling",
"url": "https://github.com/sirgraystar/mandyville-modelling/pull/38",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
375225740 | Please add gh-pages branch
Could you please create an empty gh-pages branch as target for documentation pull-requests with something like:
git checkout --orphan gh-pages
git commit --allow-empty -m "empty documentation branch: pull-request target"
git push -u origin gh-pages
I'd ideally like to see the output of godoc (and possibly a code-coverage report) end up in that branch.
In go ecosystem we already have godoc. It is a well known place where packages documentation is centralized.
Perhaps this is more of a CI/CD discussion, but my concern was ability to easily review godoc (tool) output before the resulting changes got merged into master. Similarly, i’d Like to be able to review changes in code coverage before they get merged into master. For released versions, I agree that https://godoc.org/github.com/sirupsen/logrus is the best place for that.
| gharchive/issue | 2018-10-29T21:32:16 | 2025-04-01T06:45:48.254440 | {
"authors": [
"dgsb",
"loren-osborn"
],
"repo": "sirupsen/logrus",
"url": "https://github.com/sirupsen/logrus/issues/853",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
476102071 | Writer() wraps with INFO level
What I did
log.SetLevel(log.DebugLevel)
stdoutLogger := log.New()
stdoutLogger.SetOutput(os.Stdout)
stdoutLogger.SetLevel(log.DebugLevel)
log.SetOutput(stdoutLogger.Writer())
log.Debug("asdf")
Expected behavior
Same or at least similar as after
log.SetLevel(log.DebugLevel)
log.Debug("asdf")
i.e.
DEBU[0000] asdf
Actual behavior
In reality, this results in
INFO[0000] time="2019-08-02T11:59:19+02:00" level=debug msg=asdf
I'm guessing it's something to do with the formatter?
use log.SetOutput(stdoutLogger.WriterLevel(log.DebugLevel)) instead of log.SetOutput(stdoutLogger.Writer())
See also: https://github.com/sirupsen/logrus/pull/1017
| gharchive/issue | 2019-08-02T10:04:48 | 2025-04-01T06:45:48.258558 | {
"authors": [
"c-bg",
"freeformz",
"pradykaushik"
],
"repo": "sirupsen/logrus",
"url": "https://github.com/sirupsen/logrus/issues/997",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
496591788 | Use Immutable Types
[Suggestion (2) in the series started in #14]
Usually, when a mutable object is instantiated, it allocates memory on the heap, which, in certain cases where caching is important, like inner loops or within Monte Carlo Tree Search, etc. can make them much slower. I think you could achieve significant speedup by making all of the types that you can immutable, especially things like GaussianBelief that are likely to be created and recreated often :)
All structs fixed to immutable in https://github.com/sisl/GaussianFilters.jl/commit/78c1fbc99659b4d35d3ee25575d371588d23d293
| gharchive/issue | 2019-09-21T00:26:52 | 2025-04-01T06:45:48.263149 | {
"authors": [
"jamgochiana",
"zsunberg"
],
"repo": "sisl/GaussianFilters.jl",
"url": "https://github.com/sisl/GaussianFilters.jl/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1096996039 | changed structure of 2nn estimator
As discussed with @diegodoimo, together with @imacocco I changed the inner working of the 2nn density estimator
Closing this, @diegodoimo have a look anyway
When decimation=1 the dataset must not be shuffled in compute_id_2nn
| gharchive/pull-request | 2022-01-08T17:56:14 | 2025-04-01T06:45:48.273851 | {
"authors": [
"AldoGl",
"diegodoimo"
],
"repo": "sissa-data-science/DADApy",
"url": "https://github.com/sissa-data-science/DADApy/pull/19",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2105075879 | Feature request: custom css
I found the bookmark list very confusing because the individual bookmarks are not very good visually separated. I changed this directly in the CSS so that a horizontal line appears under each bookmark, but that's not the point of editing in the code.
It would be really nice to have either an input field for a custom css classes, or the possibility of a path to a custom css file on the server. The effort for implementing the second solution suggestion should be more or less simple. The custom css should not replace the default css but override settings of the original.
The request for custom CSS often means that the design is lacking in some way. I'd be interested to know what your custom CSS looks like, so I understand better what problem, and you are solving it :).
I would not want to claim that things lacks. I love this app and I'm very happy that I could switch with this app from raindrop.io to this selfhosted app. I would like to emphasize that my impression is very, very subjective and not a critism of the design of the app. For me the design was to poppy and very confusing because the bright colors set the wrong accents. This made it difficult for me to grasp things quickly. I started with a divider line between each item, then I remember what I prefer in terms of eye guided design and raise some things. As result I changed the colors for better separate the important from the less important things, made the title bold and made also
some indentions for items
Overall this fits much better to my idea of good readable. But as written before - this is subjective. I guess there are a lot of users who are very satisfied with the default design. For my opinion individual design claims should solved by the user, not by the developer. Easiest way could be to add as last ccs link on the page which reffers to a custom css.
No plans to change the default design, but being able to add custom CSS would be neat. Could just be a plain text area in the settings where you can dump your custom CSS that would then get rendered into the HTML document. Maybe allow mounting a custom CSS file into the container, though that seems less helpful / convenient than the other option.
I would not want to claim that things lacks. I love this app and I'm very happy that I could switch with this app from raindrop.io to this selfhosted app. I would like to emphasize that my impression is very, very subjective and not a critism of the design of the app. For me the design was to poppy and very confusing because the bright colors set the wrong accents. This made it difficult for me to grasp things quickly. I started with a divider line between each item, then I remember what I prefer in terms of eye guided design and raise some things. As result I changed the colors for better separate the important from the less important things, made the title bold and made also some indentions for items
Overall this fits much better to my idea of good readable. But as written before - this is subjective. I guess there are a lot of users who are very satisfied with the default design. For my opinion individual design claims should solved by the user, not by the developer. Easiest way could be to add as last ccs link on the page which reffers to a custom css.
Hello! I like how this looks - nice work! Would you be willing to share the custom CSS you're using to make this theme?
Appreciate the custom CSS feature!
Hello! I like how this looks - nice work! Would you be willing to share the custom CSS you're using to make this theme?
Appreciate the custom CSS feature!
`a {
color: #1987e1;
outline: none;
text-decoration: none
}
a:focus,
a:hover,
a:active,
a.active {
color: #1987e1;
text-decoration: underline
}
.btn {
appearance: none;
background: #fff;
border: .05rem solid #1987e1;
border-radius: 5px;
color: #1987e1;
cursor: pointer;
display: inline-block;
font-size: .8rem;
height: 1.8rem;
line-height: 1.2rem;
outline: none;
padding: .25rem .4rem;
text-align: center;
text-decoration: none;
transition: background .2s, border .2s, box-shadow .2s, color .2s;
user-select: none;
vertical-align: middle;
white-space: nowrap
}
.btn:focus,
.btn:hover {
background: #1987e1;
border-color: #1987e1;
color: #ffffff;
text-decoration: none
}
.btn:active,
.btn.active {
background: #1987e1;
border-color: #1987e1;
color: #fff;
text-decoration: none
}
.btn.btn-primary {
background: #1987e1;
border-color: #127ad0;
color: #fff
}
.btn.btn-primary:focus,
.btn.btn-primary:hover {
background: #106dba;
border-color: #3634d2;
color: #fff
}
.btn.btn-primary:active,
.btn.btn-primary.active {
background: #127ad0;
border-color: #127ad0;
color: #fff
}
.btn.btn-success {
background: #32b643;
border-color: #2faa3f;
color: #fff
}
.btn.btn-link {
background: transparent;
border-color: transparent;
color: #1987e1
}
.btn.btn-link:focus,
.btn.btn-link:hover,
.btn.btn-link:active,
.btn.btn-link.active {
color: #1987e1
}
.btn-link:visited:not(.btn-primary) {
color: #1987e1
}
.btn-link:visited:not(.btn-primary):hover {
color: #1987e1
}
.form-input:focus {
box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2);
border-color: #1987e1
}
.form-checkbox input:focus+.form-icon,
.form-radio input:focus+.form-icon,
.form-switch input:focus+.form-icon {
box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2);
border-color: #1987e1
}
.form-checkbox input:checked+.form-icon,
.form-radio input:checked+.form-icon,
.form-switch input:checked+.form-icon {
background: #1987e1;
border-color: #1987e1
}
.form-checkbox input:indeterminate+.form-icon {
background: #1987e1;
border-color: #1987e1
}
.label.label-primary {
background: #1987e1;
color: #fff
}
.label.label-secondary {
background: #f1f1fc;
color: #1987e1
}
.badge[data-badge]::after,
.badge:not([data-badge])::after {
background: #1987e1;
background-clip: padding-box;
border-radius: .5rem;
box-shadow: 0 0 0 0.1rem #fff;
color: #fff;
content: attr(data-badge);
display: inline-block;
transform: translate(-0.05rem, -0.5rem)
}
.menu .menu-item>a:active,
.menu .menu-item>a.active {
background: #f1f1fc;
color: #1987e1
}
.pagination .page-item a:focus,
.pagination .page-item a:hover {
color: #1987e1
}
.pagination .page-item.disabled a {
cursor: default;
opacity: .5;
pointer-events: none
}
.pagination .page-item.active a {
background: #1987e1;
color: #fff
}
.tab .tab-item.active a,
.tab .tab-item a.active {
border-bottom-color: #1987e1;
color: #1987e1
}
.text-primary {
color: #1987e1 !important
}
.loading::after {
animation: loading 500ms infinite linear;
border: .1rem solid #1987e1;
border-radius: 50%;
border-right-color: transparent;
border-top-color: transparent;
content: "";
display: block;
height: .8rem;
left: 50%;
margin-left: -.4rem;
margin-top: -.4rem;
position: absolute;
top: 50%;
width: .8rem;
z-index: 1
}
.title span {
font-weight: 600;
color: #444444 !important;
}
.tag-cloud {
font-size: 0.7rem;
line-height: 1rem;
}
.group{
border-bottom: 1px #c0c0c0 solid;
padding-top: 3px;
padding-bottom: 3px;
}
.url-path, .description, .actions{
padding-left: 24px;
}
li[ld-bookmark-item] .description {
color: #888888
}
li[ld-bookmark-item] .description a,
li[ld-bookmark-item] .description a:visited:hover {
color: #b0b0b0OQ
}
li[ld-bookmark-item] {
position: relative;
border-bottom: 1px solid #c0c0c0;
padding-bottom: 10px;
}
.bookmarks-form .btn.form-icon:focus,
.bookmarks-form .btn.form-icon:hover,
.bookmarks-form .btn.form-icon:active,
.bookmarks-form .btn.form-icon.active {
color: #888888
}
li[ld-bookmark-item] .description {
color:#888888 !important;
}
li[ld-bookmark-item] .description a, li[ld-bookmark-item] .description a:visited:hover {
color:#1885de !important;
}
li[ld-bookmark-item] .url-display {
color: #b0b0b0;
}
li[ld-bookmark-item] .description {
color: #888888
}
li[ld-bookmark-item] .description a,
li[ld-bookmark-item] .description a:visited:hover {
color: #b0b0b0OQ
}
.tag-cloud .highlight-char {
font-weight: bold;
text-transform: uppercase;
color: #b0b0b0
}
.tag-cloud .selected-tags a,
.tag-cloud .selected-tags a:visited:hover {
color: #e85600
}
.tag-cloud .unselected-tags a,
.tag-cloud .unselected-tags a:visited:hover {
color: #b0b0b0
}`
Hello! I like how this looks - nice work! Would you be willing to share the custom CSS you're using to make this theme?
Appreciate the custom CSS feature!
`a { color: #1987e1; outline: none; text-decoration: none } a:focus, a:hover, a:active, a.active { color: #1987e1; text-decoration: underline }
.btn { appearance: none; background: #fff; border: .05rem solid #1987e1; border-radius: 5px; color: #1987e1; cursor: pointer; display: inline-block; font-size: .8rem; height: 1.8rem; line-height: 1.2rem; outline: none; padding: .25rem .4rem; text-align: center; text-decoration: none; transition: background .2s, border .2s, box-shadow .2s, color .2s; user-select: none; vertical-align: middle; white-space: nowrap } .btn:focus, .btn:hover { background: #1987e1; border-color: #1987e1; color: #ffffff; text-decoration: none }
.btn:active, .btn.active { background: #1987e1; border-color: #1987e1; color: #fff; text-decoration: none } .btn.btn-primary { background: #1987e1; border-color: #127ad0; color: #fff } .btn.btn-primary:focus, .btn.btn-primary:hover { background: #106dba; border-color: #3634d2; color: #fff } .btn.btn-primary:active, .btn.btn-primary.active { background: #127ad0; border-color: #127ad0; color: #fff } .btn.btn-success { background: #32b643; border-color: #2faa3f; color: #fff } .btn.btn-link { background: transparent; border-color: transparent; color: #1987e1 } .btn.btn-link:focus, .btn.btn-link:hover, .btn.btn-link:active, .btn.btn-link.active { color: #1987e1 } .btn-link:visited:not(.btn-primary) { color: #1987e1 } .btn-link:visited:not(.btn-primary):hover { color: #1987e1 }
.form-input:focus { box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2); border-color: #1987e1 }
.form-checkbox input:focus+.form-icon, .form-radio input:focus+.form-icon, .form-switch input:focus+.form-icon { box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2); border-color: #1987e1 }
.form-checkbox input:checked+.form-icon, .form-radio input:checked+.form-icon, .form-switch input:checked+.form-icon { background: #1987e1; border-color: #1987e1 } .form-checkbox input:indeterminate+.form-icon { background: #1987e1; border-color: #1987e1 }
.label.label-primary { background: #1987e1; color: #fff }
.label.label-secondary { background: #f1f1fc; color: #1987e1 }
.badge[data-badge]::after, .badge:not([data-badge])::after { background: #1987e1; background-clip: padding-box; border-radius: .5rem; box-shadow: 0 0 0 0.1rem #fff; color: #fff; content: attr(data-badge); display: inline-block; transform: translate(-0.05rem, -0.5rem) }
.menu .menu-item>a:active, .menu .menu-item>a.active { background: #f1f1fc; color: #1987e1 }
.pagination .page-item a:focus, .pagination .page-item a:hover { color: #1987e1 }
.pagination .page-item.disabled a { cursor: default; opacity: .5; pointer-events: none }
.pagination .page-item.active a { background: #1987e1; color: #fff }
.tab .tab-item.active a, .tab .tab-item a.active { border-bottom-color: #1987e1; color: #1987e1 }
.text-primary { color: #1987e1 !important }
.loading::after { animation: loading 500ms infinite linear; border: .1rem solid #1987e1; border-radius: 50%; border-right-color: transparent; border-top-color: transparent; content: ""; display: block; height: .8rem; left: 50%; margin-left: -.4rem; margin-top: -.4rem; position: absolute; top: 50%; width: .8rem; z-index: 1 } .title span { font-weight: 600; color: #444444 !important; } .tag-cloud { font-size: 0.7rem; line-height: 1rem; } .group{ border-bottom: 1px #c0c0c0 solid; padding-top: 3px; padding-bottom: 3px; } .url-path, .description, .actions{ padding-left: 24px; }
li[ld-bookmark-item] .description { color: #888888 }
li[ld-bookmark-item] .description a, li[ld-bookmark-item] .description a:visited:hover { color: #b0b0b0OQ
} li[ld-bookmark-item] { position: relative; border-bottom: 1px solid #c0c0c0; padding-bottom: 10px; }
.bookmarks-form .btn.form-icon:focus, .bookmarks-form .btn.form-icon:hover, .bookmarks-form .btn.form-icon:active, .bookmarks-form .btn.form-icon.active { color: #888888 }
li[ld-bookmark-item] .description { color:#888888 !important; }
li[ld-bookmark-item] .description a, li[ld-bookmark-item] .description a:visited:hover { color:#1885de !important; }
li[ld-bookmark-item] .url-display { color: #b0b0b0; } li[ld-bookmark-item] .description { color: #888888 } li[ld-bookmark-item] .description a, li[ld-bookmark-item] .description a:visited:hover { color: #b0b0b0OQ }
.tag-cloud .highlight-char { font-weight: bold; text-transform: uppercase; color: #b0b0b0 } .tag-cloud .selected-tags a, .tag-cloud .selected-tags a:visited:hover { color: #e85600 }
.tag-cloud .unselected-tags a, .tag-cloud .unselected-tags a:visited:hover { color: #b0b0b0 }`
请问这个代码有更新吗?它会造成搜索栏变形
Is there an update to this code please? It causes the search bar to be distorted
Hi ... here is ...
working fine with Version 1.28.0 (latest: 1.30.0 not proofed)
`:root {
--theme: #1a87e1 !important;
theme_color: #1a87e1;
}
body {
theme-color: #1a87e1 !important;
}
a {
color: #1987e1;
outline: none;
text-decoration: none
}
a:focus,
a:hover,
a:active,
a:visited,
a.active {
color: #1987e1;
text-decoration: none;
}
.btn {
appearance: none;
background: #fff;
border: .05rem solid #1987e1;
border-radius: 5px;
color: #1987e1;
cursor: pointer;
display: inline-block;
font-size: .7rem;
/* height: 1.8rem; /
/ line-height: 1.2rem; */
outline: none;
padding: .25rem .4rem;
text-align: center;
text-decoration: none;
transition: background .2s, border .2s, box-shadow .2s, color .2s;
user-select: none;
vertical-align: middle;
white-space: nowrap
}
.btn:focus,
.btn:hover {
background: #1987e1;
border-color: #1987e1;
color: #ffffff;
text-decoration: none
}
.btn:active,
.btn.active {
background: #1987e1;
border-color: #1987e1;
color: #fff;
text-decoration: none
}
.btn.btn-primary {
background: #1987e1;
border-color: #127ad0;
color: #fff
}
.btn.btn-primary:focus,
.btn.btn-primary:hover {
background: #106dba;
border-color: #3634d2;
color: #fff
}
.btn.btn-primary:active,
.btn.btn-primary.active {
background: #127ad0;
border-color: #127ad0;
color: #fff
}
.btn.btn-success {
background: #32b643;
border-color: #2faa3f;
color: #fff;
}
.btn.btn-link {
background: transparent;
border-color: transparent;
color: #1987e1;
}
.btn.btn-link:focus,
.btn.btn-link:hover,
.btn.btn-link:active,
.btn.btn-link.active {
color: #1987e1;
}
.btn-link:visited:not(.btn-primary) {
color: #1987e1 !important;
}
.btn-link:visited:not(.btn-primary):hover {
color: #1987e1;
}
.form-input{
border: .05rem solid #c0c0c0;
color: #606060;
}
.form-input:focus {
box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2);
border-color: #1987e1;
color: #606060;
}
.form-checkbox input:focus+.form-icon,
.form-radio input:focus+.form-icon,
.form-switch input:focus+.form-icon {
box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2);
border-color: #1987e1
}
.form-checkbox input:checked+.form-icon,
.form-radio input:checked+.form-icon,
.form-switch input:checked+.form-icon {
background: #1987e1;
border-color: #1987e1;
}
.form-checkbox input:indeterminate+.form-icon {
background: #1987e1;
border-color: #1987e1;
}
.label.label-primary {
background: #1987e1;
color: #fff;
}
.label.label-secondary {
background: #f1f1fc;
color: #1987e1
}
.badge[data-badge]::after,
.badge:not([data-badge])::after {
background: #1987e1;
background-clip: padding-box;
border-radius: .5rem;
box-shadow: 0 0 0 0.1rem #fff;
color: #fff;
content: attr(data-badge);
display: inline-block;
transform: translate(-0.05rem, -0.5rem)
}
.menu .menu-item>a:active,
.menu .menu-item>a.active {
background: #f1f1fc;
color: #1987e1
}
.pagination .page-item a:focus,
.pagination .page-item a:hover {
color: #1987e1
}
.pagination .page-item.disabled a {
cursor: default;
opacity: .5;
pointer-events: none
}
.pagination .page-item.active a {
background: #1987e1;
color: #fff
}
.tab .tab-item.active a,
.tab .tab-item a.active {
border-bottom-color: #1987e1;
color: #1987e1
}
.text-primary {
color: #1987e1 !important
}
.loading::after {
animation: loading 500ms infinite linear;
border: .1rem solid #1987e1;
border-radius: 50%;
border-right-color: transparent;
border-top-color: transparent;
content: "";
display: block;
height: .8rem;
left: 50%;
margin-left: -.4rem;
margin-top: -.4rem;
position: absolute;
top: 50%;
width: .8rem;
z-index: 1
}
.title span {
font-weight: 600;
color: #444444 !important;
}
.tag-cloud {
font-size: 0.7rem;
line-height: 1rem;
}
.group{
border-bottom: 1px #c0c0c0 solid;
padding-top: 3px;
padding-bottom: 3px;
}
.url-path, .description, .actions{
padding-left: 24px;
}
li[ld-bookmark-item] .description {
color: #888888
}
li[ld-bookmark-item] .description a,
li[ld-bookmark-item] .description a:visited:hover {
color: #b0b0b0OQ
}
li[ld-bookmark-item] {
position: relative;
border-bottom: 1px solid #c0c0c0;
padding-bottom: 10px;
}
.bookmarks-form .btn.form-icon:focus,
.bookmarks-form .btn.form-icon:hover,
.bookmarks-form .btn.form-icon:active,
.bookmarks-form .btn.form-icon.active {
color: #888888
}
li[ld-bookmark-item] .description {
color:#888888 !important;
}
li[ld-bookmark-item] .description a, li[ld-bookmark-item] .description a:visited:hover {
color:#1885de !important;
}
li[ld-bookmark-item] .url-display {
color: #b0b0b0;
}
li[ld-bookmark-item] .description {
color: #888888
}
li[ld-bookmark-item] .description a,
li[ld-bookmark-item] .description a:visited:hover {
color: #b0b0b0OQ
}
.tag-cloud .highlight-char {
font-weight: bold;
text-transform: uppercase;
color: #b0b0b0
}
.tag-cloud .selected-tags a,
.tag-cloud .selected-tags a:visited:hover {
color: #e85600
}
.tag-cloud .unselected-tags a,
.tag-cloud .unselected-tags a:visited:hover {
color: #b0b0b0
}
.tags a {
color: #1a87e1 !important;
}`
This is my update the custom css to the current version 1.35.0
`:root {
theme_color: #1a87e1;
--theme: #1a87e1 !important;
--primary-color: #1a87e1;
--primary-color-shade: #dfebf5;
--primary-text-color: #1a87e1;
--link-color: #1a87e1;
--gray-50: #f9f9f9;
--gray-100: #f3f3f3;
--gray-200: #e5e5e5;
--gray-300: #d1d1d1;
--gray-400: #9c9c9c;
--gray-500: #6b6b6b;
--gray-600: #4b4b4b;
--gray-700: #373737;
--gray-800: #1f1f1f;
--gray-900: #111111;
}
body {
theme-color: #1a87e1 !important;
}
a {
color: #1987e1;
outline: none;
text-decoration: none
}
a:focus,
a:hover,
a:active,
a:visited,
a.active {
color: #1987e1;
text-decoration: none;
}
.btn {
appearance: none;
background: #fff;
border: .05rem solid #1987e1;
border-radius: 5px;
color: #1987e1;
cursor: pointer;
display: inline-block;
font-size: .7rem;
/* height: 1.8rem; /
/ line-height: 1.2rem; */
outline: none;
padding: .25rem .4rem;
text-align: center;
text-decoration: none;
transition: background .2s, border .2s, box-shadow .2s, color .2s;
user-select: none;
vertical-align: middle;
white-space: nowrap
}
.btn:focus,
.btn:hover {
background: #1987e1;
border-color: #1987e1;
color: #ffffff;
text-decoration: none
}
.btn:active,
.btn.active {
background: #1987e1;
border-color: #1987e1;
color: #fff;
text-decoration: none
}
.btn.btn-primary {
background: #1987e1;
border-color: #127ad0;
color: #fff
}
.btn.btn-primary:focus,
.btn.btn-primary:hover {
background: #106dba;
border-color: #3634d2;
color: #fff
}
.btn.btn-primary:active,
.btn.btn-primary.active {
background: #127ad0;
border-color: #127ad0;
color: #fff
}
.btn.btn-success {
background: #32b643;
border-color: #2faa3f;
color: #fff;
}
.btn.btn-link {
background: transparent;
border-color: transparent;
color: #1987e1;
}
.btn.btn-link:focus,
.btn.btn-link:hover,
.btn.btn-link:active,
.btn.btn-link.active {
color: #1987e1;
}
.btn-link:visited:not(.btn-primary) {
color: #1987e1 !important;
}
.btn-link:visited:not(.btn-primary):hover {
color: #1987e1;
}
.form-input{
border: .05rem solid #c0c0c0;
color: #606060;
}
.form-input:focus {
box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2);
border-color: #1987e1;
color: #606060;
}
.form-checkbox input:focus+.form-icon,
.form-radio input:focus+.form-icon,
.form-switch input:focus+.form-icon {
box-shadow: 0 0 0 0.1rem rgba(87, 85, 217, 0.2);
border-color: #1987e1
}
.form-checkbox input:checked+.form-icon,
.form-radio input:checked+.form-icon,
.form-switch input:checked+.form-icon {
background: #1987e1;
border-color: #1987e1;
}
.form-checkbox input:indeterminate+.form-icon {
background: #1987e1;
border-color: #1987e1;
}
.label.label-primary {
background: #1987e1;
color: #fff;
}
.label.label-secondary {
background: #f1f1fc;
color: #1987e1
}
.badge[data-badge]::after,
.badge:not([data-badge])::after {
background: #1987e1;
background-clip: padding-box;
border-radius: .5rem;
box-shadow: 0 0 0 0.1rem #fff;
color: #fff;
content: attr(data-badge);
display: inline-block;
transform: translate(-0.05rem, -0.5rem)
}
.menu .menu-item>a:active,
.menu .menu-item>a.active {
background: #f1f1fc;
color: #1987e1
}
.pagination .page-item a:focus,
.pagination .page-item a:hover {
color: #1987e1
}
.pagination .page-item.disabled a {
cursor: default;
opacity: .5;
pointer-events: none
}
.pagination .page-item.active a {
background: #1987e1;
color: #fff
}
.tab .tab-item.active a,
.tab .tab-item a.active {
border-bottom-color: #1987e1;
color: #1987e1
}
.text-primary {
color: #1987e1 !important
}
.loading::after {
animation: loading 500ms infinite linear;
border: .1rem solid #1987e1;
border-radius: 50%;
border-right-color: transparent;
border-top-color: transparent;
content: "";
display: block;
height: .8rem;
left: 50%;
margin-left: -.4rem;
margin-top: -.4rem;
position: absolute;
top: 50%;
width: .8rem;
z-index: 1
}
.title span {
font-weight: 600;
color: #444444 !important;
}
.tag-cloud {
font-size: 0.7rem;
line-height: 1rem;
}
.group{
border-bottom: 1px #c0c0c0 solid;
padding-top: 3px;
padding-bottom: 3px;
}
.url-path, .description, .actions{
padding-left: 24px;
}
li[ld-bookmark-item] .description {
color: #888888
}
li[ld-bookmark-item] .description a,
li[ld-bookmark-item] .description a:visited:hover {
color: #b0b0b0OQ
}
li[ld-bookmark-item] {
position: relative;
border-bottom: 1px solid #c0c0c0;
padding-bottom: 10px;
}
.bookmarks-form .btn.form-icon:focus,
.bookmarks-form .btn.form-icon:hover,
.bookmarks-form .btn.form-icon:active,
.bookmarks-form .btn.form-icon.active {
color: #888888
}
li[ld-bookmark-item] .description {
color:#888888 !important;
}
li[ld-bookmark-item] .description a, li[ld-bookmark-item] .description a:visited:hover {
color:#1885de !important;
}
li[ld-bookmark-item] .url-display {
color: #b0b0b0;
}
li[ld-bookmark-item] .description {
color: #888888
}
li[ld-bookmark-item] .description a,
li[ld-bookmark-item] .description a:visited:hover {
color: #b0b0b0OQ
}
.tag-cloud .highlight-char {
font-weight: bold;
text-transform: uppercase;
color: #b0b0b0
}
.tag-cloud .selected-tags a,
.tag-cloud .selected-tags a:visited:hover {
color: #e85600
}
.tag-cloud .unselected-tags a,
.tag-cloud .unselected-tags a:visited:hover {
color: #b0b0b0
}
.tags a {
color: #1a87e1 !important;
}`
| gharchive/issue | 2024-01-29T09:47:02 | 2025-04-01T06:45:48.374802 | {
"authors": [
"Fischmuetze",
"GD2021",
"crosbyh",
"jonathan-s",
"sissbruecker"
],
"repo": "sissbruecker/linkding",
"url": "https://github.com/sissbruecker/linkding/issues/621",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1694964594 | Updates to BookmarkSerializer and API Documentation
This pull request includes three changes to the project.
🐛 Firstly, a bug was fixed in the BookmarkSerializer by adding missing fields such as favicon_file and web_archive_snapshot_url.
✨ Secondly, new fields is_mine and owner were added to the BookmarkSerializer to indicate whether the bookmark belongs to the authenticated user, and the name and id of the owning user is included.
📝 Finally, the shared bookmarks endpoint is now documented in the API documentation. These changes aim to improve the functionality and usability of the project.
Closing as stale
| gharchive/pull-request | 2023-05-03T23:05:05 | 2025-04-01T06:45:48.377834 | {
"authors": [
"sissbruecker",
"use-the-fork"
],
"repo": "sissbruecker/linkding",
"url": "https://github.com/sissbruecker/linkding/pull/462",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
95924442 | Coveralls coverage badge
We :heart: testable code, so adding coveralls will make it easier to track our code quality.
[x] Write tests for code written in #1
[x] Add JaCoCo support
[x] Add Coveralls plugin
[x] Integrate in Travis
[x] Add the badge to readme
Enable for PRs in the repo [Coveralls is not listed as a service]
There seems to be an issue in lemurheavy/coveralls-public#522 because of which our repo's stats aren't being taken up by coveralls. Even if we merge #9 we should wait until this issue is resolved before we close this issue.
| gharchive/issue | 2015-07-19T15:28:53 | 2025-04-01T06:45:48.411159 | {
"authors": [
"sivakumar-kailasam"
],
"repo": "sivakumar-kailasam/gradle-npm-wrapper",
"url": "https://github.com/sivakumar-kailasam/gradle-npm-wrapper/issues/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1146218708 | Can't create new tokens when using a second access token with Github sync
Describe the bug
This is a bit of a weird one, I'll try to clarify the issue once I understand the expected way this is supposed to work.
I'm having an issue with my team where when we have multiple people using Figma Tokens with GitHub sync: One of the editors can't make tokens, but only if they use their own personal access token. They have read access all tokens, but can't create new ones.
We were able to resolve this by having the second editor use my access token instead.
Is it intended that a single access token should be used to sync with GitHub? I looked through all the open issues and the documentation but didn't see a mention of this.
Things I checked before reporting:
Second editor has:
Manage access to the repo
Editor access in Figma
A fresh access token with repo access
Cleared figma cache, tried both desktop and web, on multiple machines
To Reproduce
Steps to reproduce the behavior:
Create figma file with multiple editors
Add tokens to file
Creator turns on token sync using GitHub and sets up with their credentials
Second editor sets up token sync using their own personal access token
Second editor cannot create new tokens (clicking the [+] in the UI does nothing.)
Expected behavior
I'd expect each contributor to use their own access token so that the commit history accurately reflects who made the changes, etc.
JSON (optional)
Probably not relevant, but will include JSON anyways:
{
"core": {
"color": {
"neutral": {
"0": {
"value": "#ffffff",
"type": "color"
},
"50": {
"value": "#f8f8f8",
"type": "color"
},
"100": {
"value": "#f0f0f0",
"type": "color"
},
"150": {
"value": "#e1e1e1",
"type": "color"
},
"200": {
"value": "#d2d2d2",
"type": "color"
},
"250": {
"value": "#c3c3c3",
"type": "color"
},
"300": {
"value": "#b4b4b4",
"type": "color"
},
"350": {
"value": "#a5a5a5",
"type": "color"
},
"400": {
"value": "#969696",
"type": "color"
},
"450": {
"value": "#878787",
"type": "color"
},
"500": {
"value": "#787878",
"type": "color"
},
"550": {
"value": "#696969",
"type": "color"
},
"600": {
"value": "#5a5a5a",
"type": "color"
},
"650": {
"value": "#4b4b4b",
"type": "color"
},
"700": {
"value": "#3c3c3c",
"type": "color"
},
"750": {
"value": "#262626",
"type": "color"
},
"800": {
"value": "#1e1e1e",
"type": "color"
},
"850": {
"value": "#0f0f0f",
"type": "color"
},
"900": {
"value": "#000000",
"type": "color"
}
},
"steel": {
"0": {
"value": "#F8FAFC",
"type": "color"
}
},
"gray": {
"0": {
"value": "#F9FAFB",
"type": "color"
}
},
"rose": {
"0": {
"value": "#FFF1F2",
"type": "color"
}
},
"red": {
"0": {
"value": "#FEF2F2",
"type": "color"
},
"100": {
"value": "#FEE2E2",
"type": "color"
},
"200": {
"value": "#FECACA",
"type": "color"
},
"300": {
"value": "#FCA5A5",
"type": "color"
},
"400": {
"value": "#F87171",
"type": "color"
},
"500": {
"value": "#EF4444",
"type": "color"
},
"600": {
"value": "#DC2626",
"type": "color"
},
"700": {
"value": "#B91C1C",
"type": "color"
},
"800": {
"value": "#991B1B",
"type": "color"
},
"900": {
"value": "#7F1D1D",
"type": "color"
}
},
"orange": {
"0": {
"value": "#FFF7ED",
"type": "color"
}
},
"amber": {
"0": {
"value": "#FFFBEB",
"type": "color"
}
},
"yellow": {
"0": {
"value": "#FEFCE8",
"type": "color"
}
},
"lime": {
"0": {
"value": "#F7FEE7",
"type": "color"
}
},
"green": {
"0": {
"value": "#F0FDF4",
"type": "color"
}
},
"emerald": {
"0": {
"value": "#ECFDF5",
"type": "color"
}
},
"teal": {
"0": {
"value": "#F0FDFA",
"type": "color"
}
},
"cyan": {
"0": {
"value": "#ECFEFF",
"type": "color"
}
},
"sky": {
"0": {
"value": "#F0F9FF",
"type": "color"
}
},
"blue": {
"0": {
"value": "#EFF6FF",
"type": "color"
}
},
"indigo": {
"0": {
"value": "#EEF2FF",
"type": "color"
}
},
"violet": {
"0": {
"value": "#F5F3FF",
"type": "color"
}
},
"purple": {
"0": {
"value": "#FAF5FF",
"type": "color"
}
},
"fuschia": {
"0": {
"value": "#FDF4FF",
"type": "color"
}
},
"pink": {
"0": {
"value": "#FDF2F8",
"type": "color"
}
}
},
"text": {
"family": {
"zed-mono": {
"value": "Zed Mono",
"type": "fontFamilies"
},
"zed-sans": {
"value": "Zed Sans",
"type": "fontFamilies"
}
},
"size": {
"xs": {
"value": "12",
"type": "fontSizes"
},
"sm": {
"value": "14",
"type": "fontSizes"
},
"base": {
"value": "16",
"type": "fontSizes"
},
"lg": {
"value": "18",
"type": "fontSizes"
},
"xl": {
"value": "20",
"type": "fontSizes"
}
},
"leading": {
"xs": {
"value": "12",
"type": "lineHeights"
},
"sm": {
"value": "14",
"type": "lineHeights"
},
"base": {
"value": "16",
"type": "lineHeights"
},
"lg": {
"value": "18",
"type": "lineHeights"
},
"xl": {
"value": "20",
"type": "lineHeights"
},
"2xl": {
"value": "22",
"type": "lineHeights"
}
},
"weight": {
"normal": {
"value": "Regular",
"type": "fontWeights"
},
"bold": {
"value": "Bold",
"type": "fontWeights"
}
},
"tracking": {
"default": {
"value": "0%",
"type": "letterSpacing"
}
},
"paragraphSpacing": {
"0": {
"value": "0",
"type": "paragraphSpacing"
}
},
"case": {
"default": {
"value": "none",
"type": "textCase"
}
},
"decoration": {
"none": {
"value": "none",
"type": "textDecoration"
}
},
"ui": {
"base": {
"value": {
"fontFamily": "$text.family.zed-sans",
"fontWeight": "$text.weight.normal",
"lineHeight": "$text.leading.base",
"fontSize": "$text.size.xs",
"letterSpacing": "$text.tracking.default",
"paragraphSpacing": "$text.paragraphSpacing.0",
"textCase": "$text.case.default",
"decoration": "$text.decoration.none",
"textDecoration": "$text.decoration.none"
},
"type": "typography",
"bold": {
"value": {
"fontFamily": "$text.family.zed-sans",
"fontWeight": "$text.weight.bold",
"lineHeight": "$text.leading.base",
"fontSize": "$text.size.xs",
"letterSpacing": "$text.tracking.default",
"paragraphSpacing": "$text.paragraphSpacing.0",
"textCase": "$text.case.default",
"decoration": "$text.decoration.none",
"textDecoration": "$text.decoration.none"
},
"type": "typography"
}
},
"base-bold": {
"value": {
"fontFamily": "$text.family.zed-sans",
"fontWeight": "$text.weight.bold",
"lineHeight": "$text.leading.base",
"fontSize": "$text.size.xs",
"letterSpacing": "$text.tracking.default",
"paragraphSpacing": "$text.paragraphSpacing.0",
"textCase": "$text.case.default",
"decoration": "$text.decoration.none",
"textDecoration": "$text.decoration.none"
},
"type": "typography"
}
},
"buffer": {
"sm": {
"value": {
"fontFamily": "$text.family.zed-mono",
"fontWeight": "$text.weight.normal",
"lineHeight": "$text.leading.base",
"fontSize": "$text.size.xs",
"letterSpacing": "$text.tracking.default",
"paragraphSpacing": "$text.paragraphSpacing.0",
"textCase": "$text.case.default",
"decoration": "$text.decoration.none",
"textDecoration": "$text.decoration.none"
},
"type": "typography"
},
"base": {
"value": {
"fontFamily": "$text.family.zed-mono",
"fontWeight": "$text.weight.normal",
"lineHeight": "$text.leading.2xl",
"fontSize": "$text.size.sm",
"letterSpacing": "$text.tracking.default",
"paragraphSpacing": "$text.paragraphSpacing.0",
"textCase": "$text.case.default",
"decoration": "$text.decoration.none"
},
"type": "typography"
}
}
},
"space": {
"w-4": {
"value": "$base",
"type": "spacing"
},
"w-8": {
"value": "$base * 2",
"type": "spacing"
},
"w-12": {
"value": "$base * 3",
"type": "spacing"
},
"w-16": {
"value": "$base * 4",
"type": "spacing"
}
},
"base": {
"value": "4",
"type": "spacing"
}
},
"base": {
"color": {
"semantic": {
"alert": {
"value": "$color.red.500",
"type": "color"
},
"warning": {
"value": "#FDE047",
"type": "color"
},
"info": {
"value": "#3B82F6",
"type": "color"
},
"positive": {
"value": "#4FAC63",
"type": "color"
},
"collaborator": {
"teal": {
"value": "#2DD4BF",
"type": "color"
},
"green": {
"value": "#A3E635",
"type": "color"
},
"purple": {
"value": "#7C3AED",
"type": "color"
},
"blue": {
"value": "#3B82F6",
"type": "color"
}
}
},
"system": {
"mac-os": {
"red": {
"value": "#FF5E57",
"type": "color"
},
"yellow": {
"value": "#FFBB2E",
"type": "color"
},
"green": {
"value": "#38C149",
"type": "color"
}
}
}
},
"buffer": {
"highlight": {
"active": {
"value": "$",
"type": "color"
},
"folded": {
"value": "$",
"type": "color"
},
"match": {
"value": "$",
"type": "color"
},
"active-match": {
"value": "$",
"type": "color"
},
"diff-minus": {
"value": "$",
"type": "color"
},
"diff-add": {
"value": "$",
"type": "color"
}
},
"fold-lines": {
"value": "$",
"type": "color"
}
}
},
"dark": {
"text": {
"active": {
"value": "$color.neutral.0",
"type": "color"
},
"default": {
"value": "$color.neutral.100",
"type": "color"
},
"secondary": {
"value": "$color.neutral.350",
"type": "color"
},
"muted": {
"value": "$color.neutral.550",
"type": "color"
},
"inactive": {
"value": "$color.neutral.650",
"type": "color"
},
"minimal": {
"value": "$color.neutral.750",
"type": "color"
},
"semantic": {
"error": {
"value": "#E06C75",
"type": "color"
},
"warning": {
"value": "#D1B366",
"type": "color"
}
}
},
"surface": {
"100": {
"value": "$color.neutral.750",
"type": "color"
},
"300": {
"value": "$color.neutral.800",
"type": "color"
},
"400": {
"value": "$color.neutral.850",
"type": "color"
},
"500": {
"value": "$color.neutral.900",
"type": "color"
}
},
"border": {
"default": {
"value": "$color.neutral.850",
"type": "color"
},
"muted": {
"value": "$color.neutral.700",
"type": "color"
},
"active": {
"value": "$color.neutral.500",
"type": "color"
}
},
"icon": {
"active": {
"value": "$text.active",
"type": "color"
},
"secondary": {
"value": "$text.secondary",
"type": "color"
},
"muted": {
"value": "$text.muted",
"type": "color"
},
"inactive": {
"value": "$text.inactive",
"type": "color"
}
},
"syntax": {
"color": {
"primary": {
"value": "$text.default",
"type": "color"
},
"constant": {
"value": "$syntax.color.primary",
"type": "color"
},
"comment": {
"value": "$text.muted",
"type": "color"
},
"predictive": {
"value": "$text.muted",
"type": "color"
},
"punctuation": {
"value": "$text.secondary",
"type": "color"
},
"keyword": {
"value": "#7DD3FC",
"type": "color"
},
"function": {
"value": "#FDE68A",
"type": "color"
},
"string": {
"value": "#818CF8",
"type": "color"
},
"type": {
"value": "#6EE7B7",
"type": "color"
},
"variant": {
"value": "$syntax.color.type",
"type": "color"
},
"number": {
"value": "#FBBF24",
"type": "color"
},
"boolean": {
"value": "$syntax.color.number",
"type": "color"
},
"property": {
"value": "$syntax.color.keyword",
"type": "color"
},
"operator": {
"value": "#E879F9",
"type": "color"
},
"enum": {
"value": "$syntax.color.operator",
"type": "color"
}
},
"buffer": {
"bg": {
"value": "$surface.500",
"type": "color"
},
"selection": {
"value": "#FFFFFF14",
"type": "color"
}
},
"gutter": {
"active": {
"value": "$text.active",
"type": "color"
},
"inactive": {
"value": "$text.muted",
"type": "color"
}
}
}
},
"light": {
"text": {
"active": {
"value": "$color.neutral.900",
"type": "color"
},
"default": {
"value": "$color.neutral.800",
"type": "color"
},
"secondary": {
"value": "$color.neutral.600",
"type": "color"
},
"muted": {
"value": "$color.neutral.500",
"type": "color"
},
"inactive": {
"value": "$color.neutral.300",
"type": "color"
},
"minimal": {
"value": "$color.neutral.150",
"type": "color"
},
"semantic": {
"error": {
"value": "#E06C75",
"type": "color"
},
"warning": {
"value": "#D1B366",
"type": "color"
}
}
},
"surface": {
"100": {
"value": "$color.neutral.250",
"type": "color"
},
"300": {
"value": "$color.neutral.150",
"type": "color"
},
"400": {
"value": "$color.neutral.100",
"type": "color"
},
"500": {
"value": "$color.neutral.50",
"type": "color"
}
},
"border": {
"default": {
"value": "$color.neutral.300",
"type": "color"
},
"muted": {
"value": "$color.neutral.200",
"type": "color"
},
"active": {
"value": "$color.neutral.300",
"type": "color"
}
},
"icon": {
"active": {
"value": "$text.active",
"type": "color"
},
"secondary": {
"value": "$text.secondary",
"type": "color"
},
"muted": {
"value": "$text.muted",
"type": "color"
},
"inactive": {
"value": "$text.inactive",
"type": "color"
}
},
"syntax": {
"color": {
"primary": {
"value": "$text.default",
"type": "color"
},
"constant": {
"value": "$syntax.color.primary",
"type": "color"
},
"comment": {
"value": "$text.muted",
"type": "color"
},
"predictive": {
"value": "$text.muted",
"type": "color"
},
"punctuation": {
"value": "$text.secondary",
"type": "color"
},
"keyword": {
"value": "#7DD3FC",
"type": "color"
},
"function": {
"value": "#FDE68A",
"type": "color"
},
"string": {
"value": "#818CF8",
"type": "color"
},
"type": {
"value": "#6EE7B7",
"type": "color"
},
"variant": {
"value": "$syntax.color.type",
"type": "color"
},
"number": {
"value": "#FBBF24",
"type": "color"
},
"boolean": {
"value": "$syntax.color.number",
"type": "color"
},
"property": {
"value": "$syntax.color.keyword",
"type": "color"
},
"operator": {
"value": "#E879F9",
"type": "color"
},
"enum": {
"value": "$syntax.color.operator",
"type": "color"
}
},
"buffer": {
"bg": {
"value": "$surface.500",
"type": "color"
},
"selection": {
"value": "#0000000A",
"type": "color"
}
},
"gutter": {
"active": {
"value": "$text.active",
"type": "color"
},
"inactive": {
"value": "$text.muted",
"type": "color"
}
}
}
}
}
Are you sure the second access token has write scope defined and has access to that repository? It reads like it only has read scope defined.
I just checked this with my colleague and it works as intended, we can both pull from and push to our GitHub repo using our own access token (2 different ones, that is).
@iamnbutler My guess is your second token only has read access or if you're in an org repo you forgot to SSO enable the second token (happens a lot in our org 😄 )
Hey @iamnbutler ,
Sorry, maybe I was to quick 🙈
I see now according to your "steps to reproduce", that it might not be a GitHub sync issue, but more related to the UI in the plugin as such.
I did try to reproduce your issue, and while I can create a new token it can end up at the bottom if I have "top-level" tokens as well as nested ones.
Maybe this screenshot describes it better:
If I add a new token to "gray" using the + icon it ends up at the end of "gray" (the green circle)
If I add a new token to the top level "color" (also using the + icon) it ends up at the end of the entire "color" group and after "gray" (the red circle)
Is your issue similar or something else?
Ok great! Let me to some more through testing, and ensure we aren't having read only issues.
My guess is your second token only has read access or if you're in an org repo you forgot to SSO enable the second token (happens a lot in our org 😄 )
Oh, I don't think I've checked this. It could be the issue. Will report back!
Is your issue similar or something else?
I'll check following your steps to see if I have the same issue once resolving the above!
Were you able to resolve this @iamnbutler ? Let me know if there's anything I can do to help! 🙏
Been a bit slammed but will test this in the afternoon and see if we can't resolved!
So we spent some time on this, I'm not sure which of the things we did fixed it but my co-worker was able to write tokens after manually adding him to the repo with write access (even though he should have already had write access from the org-level settings.)
As @iamnbutler found, this has solved the same issue for our team in Gitlab.
We had two team members with owner permissions for the repo inherited from their org/team association.
We went into the project this repo belonged to and explicitly added them again (they will show up twice in the members list) as a "Direct Member." They are now able to access, pull, push, write using personal access tokens.
Thank you!
Closing this since it seems to have been solved.
| gharchive/issue | 2022-02-21T20:55:06 | 2025-04-01T06:45:48.441105 | {
"authors": [
"iamnbutler",
"jakobe",
"mtopping",
"six7"
],
"repo": "six7/figma-tokens",
"url": "https://github.com/six7/figma-tokens/issues/479",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1758955345 | I don't seem to be able to run the code correctly
Hello, I have the following problem when running the code and I need your help very much!
After I clone the code and install the dependencies according to the tutorial (i.e. complete the following steps), I open jupyternotebook and then how to execute the code?
I tried to run directly in vscode python . /datagen/gen_data.py, but the following error is reported
After encountering the above error, I tried to create an hdf5 file using the command: touch /xxx/test/deep-route/data/layout_data.hdf5, after executing python . /datagen/gen_data.py still reported the same error, and I created the hdf5 file disappeared.
I've tried many things and still can't solve the problem, hope to get your help!
There was an issue with the genData function. An issue related with the directory
The rectified function code is as below:
###########################################
def genData(N=10, H=32, W=32, pin_range=(2, 6)):
"""
Generates decoded training image dataset of size (N, C, H, W)
Inputs:
- N: number of images to generate
- H, W: image height, width (in px)
- pin_range: tuple (low, high) for allowed range of number of pins; "half-open" interval [low, high)
(e.g. (2, 6) pins means 2 or 3 or 4 or 5 pins)
Outputs:
- Saves X: training data (N, 1, H, W)
- Saves Y: training labels (N, C, H, W)
where C = 8 (each corresponds to a layout layer, viz.
[pin, m3, via3, m4, via4, m5, via5, m6])
X[:, 0, :, :] pin (m2)
Y[:, 0, :, :] pin (m2) - same as X[:, 0, :, :]
Y[:, 1, :, :] m3 (vert)
Y[:, 2, :, :] via3
Y[:, 3, :, :] m4 (horz)
Y[:, 4, :, :] via4
Y[:, 5, :, :] m5 (vert)
Y[:, 6, :, :] via5
Y[:, 7, :, :] m6 (horz)
"""
# 8 layout layers for now
C = 8
data_dir = os.getcwd() + "/data/"
# Ensure the directory exists
if not os.path.exists(data_dir):
os.makedirs(data_dir)
if os.path.exists(data_dir):
for file in os.listdir(data_dir):
if file.endswith(".hdf5"):
os.remove(data_dir+file)
else:
os.makedirs(data_dir)
data = h5py.File(data_dir + "layout_data.hdf5", 'w')
# numpy arrays no longer needed; use HDF5 instead
#X = np.zeros([N, 1, H, W], dtype = np.int8)
#Y = np.zeros([N, C, H, W], dtype = np.int8)
X = data.create_dataset("X", shape=(N, 1, H, W), dtype='uint8', compression='lzf', chunks=(1, 1, H, W))
Y = data.create_dataset("Y", shape=(N, C, H, W), dtype='uint8', compression='lzf', chunks=(1, 1, H, W))
# Set physical size represented by HxW pixels
microns = 11.0 # To have balanced dataset covering from m3 to m6 (based on resistance plots from resistance_vs_distance.ipynb)
microns_per_xpixel = microns/W
microns_per_ypixel = microns/H
# Layer map
l_map = {
# Pins
'pin' : 0,
# Vias
'via3' : 2,
'via4' : 4,
'via5' : 6,
# Vertical tracks
'm3' : 1,
'm5' : 5,
# Horizontal tracks
'm4' : 3,
'm6' : 7
}
#m3_m4 = m5_m4 = m5_m6 = 0
n = 0
print_every = 5000
while n < N:
# Randomly select number of pins from given range
# Uniform distribution over pin range
nPins = np.random.randint(*pin_range)
# Non-uniform distribution (skewed exponentially towards smaller number of pins)
#p_range = np.array(range(*pin_range))
#p = np.exp(-p_range) / np.sum(np.exp(-p_range))
#nPins = np.random.choice(p_range, p=p)
# Randomly pick x and y co-ords for nPins from [0, W) and [0, H) pixels
x_pins = np.random.randint(W, size=nPins)
y_pins = np.random.randint(H, size=nPins)
max_xlen = (max(x_pins) - min(x_pins)) * microns_per_xpixel # length in um
max_ylen = (max(y_pins) - min(y_pins)) * microns_per_ypixel # length in um
# Corner case when pins overlap each other (invalid case)
# Bug fix for https://github.com/sjain-stanford/RouteAI/issues/4
if (max_xlen == 0) and (max_ylen == 0):
continue
# Draw pins on layer 'pin (m2)' of both X (data) and Y (labels)
X[n] = drawPins(X[n], x_pins, y_pins, l_map)
Y[n] = drawPins(Y[n], x_pins, y_pins, l_map)
# Add routes to Y (labels)
x_wire, y_wire = selectWireClass(max_xlen, max_ylen)
Y[n] = drawRoutes(Y[n], x_pins, y_pins, max_xlen, max_ylen, x_wire, y_wire, l_map)
n += 1
if (n % print_every == 0):
print("Finished generating %d samples." %(n))
#if x_wire == 'm4' and y_wire == 'm3':
# m3_m4 += 1
#elif x_wire == 'm4' and y_wire == 'm5':
# m5_m4 += 1
#elif x_wire == 'm6' and y_wire == 'm5':
# m5_m6 += 1
#else:
# print(x_wire, y_wire)
#print(m3_m4, m5_m4, m5_m6)
# Storing as .npy using np.save -> Issue: RAM out of memory, disk memory limitation
#data_dir = os.getcwd() + '/data/'
#if os.path.exists(data_dir):
# for file in os.listdir(data_dir):
# if file.endswith(".npy"):
# os.remove(data_dir+file)
#else:
# os.makedirs(data_dir)
#X_save = data_dir + 'X_save.npy'
#Y_save = data_dir + 'Y_save.npy'
#np.save(X_save, X, allow_pickle=False)
#np.save(Y_save, Y, allow_pickle=False)
print("Dataset generated as follows:")
for ds in data:
print(ds, data[ds])
#############################################
| gharchive/issue | 2023-06-15T14:24:52 | 2025-04-01T06:45:48.498475 | {
"authors": [
"PandaK404",
"emillal"
],
"repo": "sjain-stanford/deep-route",
"url": "https://github.com/sjain-stanford/deep-route/issues/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
257786825 | [Redesign] Melhorias de experiência do usuário
Estive olhando o site de presença dos vereadores.
Pensei em uma abordagem de experiência do usuário melhorada pra ele.
Estou partindo do princípio que:
O usuário é um cidadão qualquer de são josé;
O usuário quer saber quais vereadores mais faltam;
O usuário quer saber se o vereador que ele acompanha falta;
Para esse redesign pensei:
Um site focado em mobile com duas abas: Vereadores e Sessões.
Aba Vereadores:
Exibe uma tela por vez;
Cada tela exibe um gráfico estilo github - cada quadrado é uma sessão;
Se o quadrado tá verdinho ele foi
Se o quadrado tá vermelho ele faltou
Clicando em uma sessão abre um modal
Esse modal tem dados do vereador e da sessão e um botão "Faltou porquê?"
Clicando em "Faltou porque" abre um textarea para escrever um texto pro vereador perguntando pq ele faltou. (vai pro email dele)
Aba Sessões:
Só pensei em uma lista de sessões, mas provavelmente dá para melhorar isso;
Outras ideias:
Uma aba "ranking" mostrando quem falta mais;
Opção de seguir um vereador e ficar sabendo se ele faltar - dá para fazer com services workers (somente javascript)
@luizfelipejr O que você acha das proposições?
Olha o sistema atual: https://sjcdigital.github.io/presenca-vereadores-sjc/
| gharchive/issue | 2017-09-14T16:49:06 | 2025-04-01T06:45:48.506597 | {
"authors": [
"PauloLuan",
"teles"
],
"repo": "sjcdigital/presenca-vereadores-sjc",
"url": "https://github.com/sjcdigital/presenca-vereadores-sjc/issues/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1831379874 | プラグインテンプレートにテストとかを追加する
やっぱテストいるやん。
元々はテスト種別を各々好きにっていうスタンスだったけどテンプレだと強制してもいい気がしてきた。
いま作ってる非参考実装プラグインの結果を適用していきたい。
| gharchive/issue | 2023-08-01T14:11:13 | 2025-04-01T06:45:48.519665 | {
"authors": [
"sk-0520"
],
"repo": "sk-0520/Pe",
"url": "https://github.com/sk-0520/Pe/issues/891",
"license": "WTFPL",
"license_type": "permissive",
"license_source": "github-api"
} |
953815113 | Enable the device server push events manually
Problem Description
At the moment the generated device server requires that polling be set for attribute(s) so that the client(s) can be able to subscribe to the attribute events. Depending on how many attributes the device might have, this might also degrade the performance of the server.
One proposed solution to this would be to have the device server push events to clients, which is more efficient than polling.
Two possible solutions (there can be more) are proposed:
Solution 1. Specify the flag(s) as part of the attribute configuration
In this scenario we add a property/properties in each attribute configuration to flag whether we want the device server to push either change/archive event for each individual attribute. The nice thing about this approach is that we can set it up for a specific attribute.
The proper place for this would be to use the SimDD.json configuration file to add this extra properties. This means we might also have to update the parser, but it shouldn't be difficult, and then when initialising the device, we would get these properties from the model and configure it accordingly.
Code snippet: tango_sim_generator.py
def initialize_dynamic_attributes(self):
model_sim_quants = self.model.sim_quantities
attribute_list = set([attr for attr in model_sim_quants.keys()])
for attribute_name in attribute_list:
meta_data = model_sim_quants[attribute_name].meta
# Dynamically add all attributes except those with DevEnum data type,
# and SPECTRUM data format since they are added statically to the device
# class prior to start-up. Also exclude attributes with a data format
# 'IMAGE' as we currently do not handle them.
if not self._is_attribute_addable_dynamically(meta_data):
if model_sim_quantities[attribute_name].meta['push_change_event']:
self.set_change_event(attribute_name, True, True)
if model_sim_quantities[attribute_name].meta['push_archive_event']:
self.set_change_event(attribute_name, True, True)
continue
....
self._add_dynamic_attribute(attr, rw_type)
MODULE_LOGGER.debug("Added dynamic {} attribute".format(attribute_name))
if model_sim_quantities[attribute_name].meta['push_change_event']:
self.set_change_event(attribute_name, True, True)
if model_sim_quantities[attribute_name].meta['push_archive_event']:
self.set_change_event(attribute_name, True, True)
Solution 2. Pass the flags as command line arguments
In this scenario, we pass a flag(s) in the terminal to signal whether to manually push events or not. With this approach, there won't be an easier way to specify as to for which attribute we want the event to be pushed for. We will have to just push for all attributes.
Code snippet: bash
$ tango-simlib-generator --sim-data-file <xmi/fgo/json>\
--dserver-name <server-name>\
--directory <output filepath>\
--push-change-events <True/False>\
--push-archive-events <True/False>
Code snippet: tango_sim_generator.py
....
def generate_device_server(
server_name,
sim_data_files,
directory="",
push_change_events=False,
push_archive_events=False
):
....
lines = [
....
" TangoDeviceServers = get_tango_device_server("
" models,"
" sim_data_files,"
" push_change_events={},".format(push_change_events)
" push_archive_events={}".format(push_archive_events)
" )"
....
def get_tango_device_server(
models,
sim_data_files,
push_change_events=False,
push_archive_events=False
):
....
def initialize_dynamic_attributes(self):
model_sim_quants = self.model.sim_quantities
attribute_list = set([attr for attr in model_sim_quants.keys()])
for attribute_name in attribute_list:
meta_data = model_sim_quants[attribute_name].meta
# Dynamically add all attributes except those with DevEnum data type,
# and SPECTRUM data format since they are added statically to the device
# class prior to start-up. Also exclude attributes with a data format
# 'IMAGE' as we currently do not handle them.
if not self._is_attribute_addable_dynamically(meta_data):
if push_archive_events:
self.set_archive_event(attribute_name, True, True)
if push_change_events:
self.set_change_event(attribute_name, True, True)
continue
....
self._add_dynamic_attribute(attr, rw_type)
MODULE_LOGGER.debug("Added dynamic {} attribute".format(attribute_name))
if push_archive_events:
self.set_archive_event(attribute_name, True, True)
if push_archive_event:
self.set_change_event(attribute_name, True, True)
@ajoubertza , @samueltwum1 please let me know what you think of this? Any suggestions are also welcome.
I like option 1, as the behaviour is more clearly document in the configuration file, and it gives the user more options. A few suggestions:
I agree that the default behaviour is not to push change or archive events.
For devices that have many attributes it will be tedious to add the config for each one to simDD.json, so I suggest a global option as the start of the file that allows us to enable the feature(s) for all attributes. Then users could optionally disable it for some attributes, if they like. Alternatively, the user could disable at the global level, and then only turn it on for a few attributes. It depends on use case.
It would also be useful to override the last parameter to set_archive_event and set_change_event. To use rel/abs filtering rules or not.
| gharchive/issue | 2021-07-27T12:08:28 | 2025-04-01T06:45:48.539196 | {
"authors": [
"ajoubertza",
"kmadisa"
],
"repo": "ska-sa/tango-simlib",
"url": "https://github.com/ska-sa/tango-simlib/issues/140",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1362472246 | Reported errors don't agree with TypeScript IDE / CLI
Describe the bug
In the above you can see that the reported error is not reported in the IDE and I know it to be correct (e.g. this would pass in Jest before etc.).
Reproduction
Run tests on this branch https://github.com/jasonkuhrt/alge/tree/tests/try-vite-ts-plugin
System Info
Node 18, vitest 0.23
Used Package Manager
pnpm
Validations
[X] Follow our Code of Conduct
[X] Read the Contributing Guide.
[X] Check that there isn't already an issue that reports the same bug to avoid creating a duplicate.
[X] Check that this is a concrete bug. For Q&A, please open a GitHub Discussion instead.
[X] The provided reproduction is a minimal reproducible of the bug.
@jasonkuhrt This is the expected behaviour, expectType is doing a strict type equality comparison and in this case the input parameter in your create function is not a simple Record but an intersection of several Zod objects. In this case, if your goal is not to test the type equality but its assignability you should use expectAssignable.
@jasonkuhrt Maybe you want to test the tssert API which is a bit more flexible and less ambiguous.
@skarab42 Do you agree it is incorrect that the IDE and Vitest output disagree though?
@jasonkuhrt I don't see any errors in the IDE except the warning that says to avoid using unbound methods.
Sorry I misunderstood the question. It's normal not to have the error in the IDE nor with tsd. If you look at the implementation of the expectType function you see that it does strictly nothing and accepts all types/values. When the plugin encounters this function then it calls a custom version of the TS compiler to perform the type tests.
@skarab42 Ahhhh. Ok did not expect that. Do you prefer to work this way? Does it have to be this way? Is the idea to centralize all the "feedback" into one consistent place?
Not sure what I think yet.
@jasonkuhrt I really need to add more documentation to this project. Ideally I would like to use tsc directly but unfortunately I haven't found a way to do it and I don't think it's possible (but you never know until someone does). However, it is possible to make a plugin for vscode that could report the errors in the IDE (not in the near future). In any case, thank you for reporting.
| gharchive/issue | 2022-09-05T21:13:14 | 2025-04-01T06:45:48.550540 | {
"authors": [
"jasonkuhrt",
"skarab42"
],
"repo": "skarab42/vite-plugin-vitest-typescript-assert",
"url": "https://github.com/skarab42/vite-plugin-vitest-typescript-assert/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
85403829 | attributes, setters and getters
It would be nice to extend logic of custom attributes to element setters / getters similar to x-tags, for example below code would allow execute custom set / get for element.overlay and propagate change to attribute
accessors: {
overlay: {
attribute: {boolean: true},
},
escapeHide: {
attribute: {
boolean: true,
name: 'escape-hide'
}
},
overlayTapHide: {
attribute: {
boolean: true,
name: 'overlay-tap-hide'
}
}
}
Hey @dredzone I've pushed this to master. You can find some details in the "properties" option in the massive options object that is documented near the top of the readme. It may change depending on feedback internal here at Atlassian as well as feedback from you guys, but what is there should be close to what will be in the next release.
Basically to achieve what you want all you'll need to do:
properties: {
myProperty: {
attr: true,
type: Boolean
}
}
thanks for quick turnaround.
| gharchive/issue | 2015-06-05T03:32:30 | 2025-04-01T06:45:48.553137 | {
"authors": [
"dredzone",
"treshugart"
],
"repo": "skatejs/skatejs",
"url": "https://github.com/skatejs/skatejs/issues/200",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1618133685 | LightSwitch
Current Behavior
export let width = "w-12";
export let height = "w-6";
Steps To Reproduce
export let width = "w-12";
export let height = "h-6";
Anything else?
No response
Hey @4ccoun7 - already found, fixed, and ready for the next release. Look for that to drop this weekend as part of our v1.0 launch event.
| gharchive/issue | 2023-03-09T23:24:08 | 2025-04-01T06:45:48.557707 | {
"authors": [
"4ccoun7",
"endigo9740"
],
"repo": "skeletonlabs/skeleton",
"url": "https://github.com/skeletonlabs/skeleton/issues/1123",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
43171917 | filtering out spurious mouseMove preventing mouseClick in Chrome
There is a problem with the current version of Chrome:
It very often emits mouseMouve event after the mouseDown event and before the mouseUp but with no user mouse movement (big killing bug!).
In this case, this should restore the click functionality, while being transparent for browser with no e.movementX support.
onDocumentMouseMove: function(e) {
Cursor.refresh(e);
if (this.dragging) {
if( !this.activity && e.movementX === 0 && e.movementY === 0 ) return;
this.activity = true;
}
},
I wouldn't optimize for Chrome bugs. Can you please let us know if this is still an issue?
| gharchive/issue | 2014-09-18T19:56:29 | 2025-04-01T06:45:48.563940 | {
"authors": [
"g-van-vreckem",
"skidding"
],
"repo": "skidding/dragdealer",
"url": "https://github.com/skidding/dragdealer/issues/58",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
579822336 | which measure of proportionality is better?
Dear Michael,
I am conducting an analysis in which I aim to rank all known sources of variance (ie cell type, donor, technical artifacts) in my single-cell RNA-seq dataset. Among others, I am computing all pairwise cell-cell distances, getting a distance matrix as an output. Your article "Evaluating measures of association for single-cell transcriptomics" has been extremely useful in this regard. I also observe a greater signal-to-noise ratio and overall accuracy when using measures of proportionality (phi and rho) as compared to Pearson correlation (as you report in figure 4).
My question is: which measure of proportionality would you use? I like rho because its bounded between [-1,1]. However, I get a great deal of negative values (ie -0.1) which I find it hard to interpret. On the other hand, phi is always positive, but is unbounded.
Thanks a lot for your time and help, and for creating this awesome package.
Best,
Ramon
Hey Ramon, sorry for the delay. Glad to hear our paper was useful to you, and that you are seeing similar results. I tend to use rho because, as you say, it’s often useful to have a measure bounded by [-1, 1]. In practice, depending on the application I’m not sure the choice is that significant; the two are related by a monotonic function and correspondingly, the differences we saw between them in our paper were quite minor. You might want to take a look at the propr paper (https://www.nature.com/articles/s41598-017-16520-0), which is the implementation that dismay is providing a fairly shallow wrapper around, for more details - the SI appendix of this paper might be particularly useful.
Hope this helps.
Mike
Thanks a lot Mike, this is very useful!
Ramon
| gharchive/issue | 2020-03-12T10:01:19 | 2025-04-01T06:45:48.582713 | {
"authors": [
"massonix",
"skinnider"
],
"repo": "skinnider/dismay",
"url": "https://github.com/skinnider/dismay/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1537666862 | [CONNECTOR] Add hdfs server, hdfsFileSystem and tests
#1391
當前方式是透過 hadoop.minicluster 創建本地端的 hdfs ,也順便寫了 hdfs 版的 sink task 測試來看看是否可以正確執行。
這邊想要請教一下關於之前 build failed 的事情
我發現如果在 module it 的 dependency 之中加入 hadoop-common,就會導致 ExporterTest > testRunWithDefaultConfigs() 與 PerfSourceTest > testDefaultConfig() 無法通過測試,而結果是 ignored,好像是因為 SERVICE.workerUrl() 所導致的
有測試過單純加入 dependency 並無增減任何程式碼也會導致此錯誤,所以在想會不會是這兩邊有衝突到。
謝謝
有測試過單純加入 dependency 並無增減任何程式碼也會導致此錯誤,所以在想會不會是這兩邊有衝突到。
應該是有相依性的衝突,例如 hadoop 仰賴 A 函式庫的0.1版、kafka worker 仰賴 A 函式庫的0.2版,這導致做整合測試的時候會出問題,可以試著用 ./gradlew it:dependencies 查看一下是否有出現互相覆蓋的函式庫,如果有的話則看一下版本之間是不是落差很大
不太確定在 hadoop-common 上設定 transitive 為 false 這樣妥不妥當。
以下是觀察到 connect-runtime 與 hadoop-common 之中有衝突的項目
connect-runtime
| ch.qos.reload4j:reload4j:1.2.19 -> 1.2.22
| ch.qos.reload4j:reload4j:1.2.19 -> 1.2.22
| jakarta.activation:jakarta.activation-api:1.2.1 -> 1.2.2
| ch.qos.reload4j:reload4j:1.2.19 -> 1.2.22
| org.javassist:javassist:3.25.0-GA -> 3.26.0-GA
| org.apache.commons:commons-lang3:3.8.1 -> 3.12.0
hadoop-common
| +--- commons-cli:commons-cli:1.2 -> 1.4
| | \--- commons-codec:commons-codec:1.11 -> 1.15
| +--- commons-io:commons-io:2.8.0 -> 2.11.0
| +--- jakarta.activation:jakarta.activation-api:1.2.1 -> 1.2.2
| +--- org.eclipse.jetty:jetty-server:9.4.43.v20210629 -> 9.4.48.v20220622 (*)
| +--- org.eclipse.jetty:jetty-util:9.4.43.v20210629 -> 9.4.48.v20220622
| +--- org.eclipse.jetty:jetty-servlet:9.4.43.v20210629 -> 9.4.48.v20220622 (*)
| | | \--- org.eclipse.jetty:jetty-util:9.4.43.v20210629 -> 9.4.48.v20220622
| | \--- org.eclipse.jetty:jetty-servlet:9.4.43.v20210629 -> 9.4.48.v20220622 (*)
| | | \--- javax.xml.bind:jaxb-api:2.2.2 -> 2.3.0
| | +--- org.codehaus.jackson:jackson-core-asl:1.9.2 -> 1.9.13
| | +--- org.codehaus.jackson:jackson-mapper-asl:1.9.2 -> 1.9.13
| | | +--- org.codehaus.jackson:jackson-core-asl:1.9.2 -> 1.9.13
| | | \--- org.codehaus.jackson:jackson-mapper-asl:1.9.2 -> 1.9.13 (*)
| | | +--- org.codehaus.jackson:jackson-core-asl:1.9.2 -> 1.9.13
| | | \--- org.codehaus.jackson:jackson-mapper-asl:1.9.2 -> 1.9.13 (*)
| +--- commons-logging:commons-logging:1.1.3 -> 1.2
| | \--- org.apache.commons:commons-lang3:3.7 -> 3.12.0
| | +--- com.thoughtworks.paranamer:paranamer:2.3 -> 2.8
| | +--- org.xerial.snappy:snappy-java:1.0.5 -> 1.1.8.4
| | +--- org.apache.commons:commons-compress:1.4.1 -> 1.21
| | \--- org.slf4j:slf4j-api:1.6.4 -> 1.7.36
| +--- com.google.protobuf:protobuf-java:2.5.0 -> 3.19.4
| | +--- org.apache.zookeeper:zookeeper:3.5.6 -> 3.8.0 (*)
| | | | | \--- org.slf4j:slf4j-api:1.7.25 -> 1.7.36
| | | | | \--- org.slf4j:slf4j-api:1.7.25 -> 1.7.36
| | | | | \--- commons-io:commons-io:2.5 -> 2.11.0
| | | | \--- com.nimbusds:nimbus-jose-jwt:3.10 -> 9.8.1 (*)
| +--- org.apache.zookeeper:zookeeper:3.5.6 -> 3.8.0 (*)
| +--- com.fasterxml.jackson.core:jackson-databind:2.12.7 -> 2.13.3 (*)
| | \--- org.codehaus.woodstox:stax2-api:4.2 -> 4.2.1
| \--- org.xerial.snappy:snappy-java:1.1.8.2 -> 1.1.8.4
@Haser0305 這是一個很好的練習,我給你一些線索,你試著查查看
首先,我們可以用指令列出明確的 error stack
./gradlew clean connector:test --tests ExporterTest --info
--info 可以秀出更多資訊
接著你可以看到大概是哪個套件出現問題,然後用該套件去 hadoop and kafka 下去找,看看是不是有明顯的版本落差(例如1版和2版)
| gharchive/pull-request | 2023-01-18T09:25:10 | 2025-04-01T06:45:48.588237 | {
"authors": [
"Haser0305",
"chia7712"
],
"repo": "skiptests/astraea",
"url": "https://github.com/skiptests/astraea/pull/1447",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
119577388 | Fix polylog eval
See this thread.
@asmeurer, thank you for using 8ee62cb in sympy/sympy#10799
| gharchive/pull-request | 2015-11-30T20:45:18 | 2025-04-01T06:45:48.590209 | {
"authors": [
"skirpichev"
],
"repo": "skirpichev/omg",
"url": "https://github.com/skirpichev/omg/pull/137",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1680158310 | Add volumemute command
With this addition, CATT should now have the ability to toggle between mute states while automatically preserving the previous volume level. This function will also prevent a volume preview tone on some devices.
Example: catt -d <DEVICE_OR_IP> volumemute <true | false>
Note that the default value is true.
Does this work for all devices?
Not exactly. I also have a Lenovo Smart Display. While that device mutes and unmutes as expected, it unfortunately still emits a volume tone.
I only have the two device models, so feedback on which brands are/are not working would be great. I suspect it to be the same as Home Assistant's media_player.volume_mute service.
Thank you! I'll review shortly.
Removed attribution in the docstring- anything else look off?
| gharchive/pull-request | 2023-04-23T20:29:56 | 2025-04-01T06:45:48.597341 | {
"authors": [
"neurodiv-eric",
"skorokithakis"
],
"repo": "skorokithakis/catt",
"url": "https://github.com/skorokithakis/catt/pull/427",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
335227334 | Ignore unused args parameter for main method
Added a special case to handle https://github.com/sksamuel/scapegoat/issues/105
:+1:
| gharchive/pull-request | 2018-06-25T00:39:12 | 2025-04-01T06:45:48.623225 | {
"authors": [
"sksamuel",
"zgrannan"
],
"repo": "sksamuel/scapegoat",
"url": "https://github.com/sksamuel/scapegoat/pull/203",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
583012281 | Sync tabs in split view
An interesting idea popped up on our.
https://our.umbraco.com/forum/umbraco-8/93641-discussion-about-tabs#comment-317850
Syncing the tabs when working in split views. I like the idea, but should probably be togglable.
If anyone wants to give it a go, I've marked it up for grabs :)
I've started working on this. I'll try to have a PR by the end of the weekend.
I created the functionality including the syncing of the scrollto functionality
PR can be found here: https://github.com/skttl/umbraco-matryoshka/pull/34
| gharchive/issue | 2020-03-17T13:27:22 | 2025-04-01T06:45:48.627831 | {
"authors": [
"FransdeJong",
"skttl"
],
"repo": "skttl/umbraco-matryoshka",
"url": "https://github.com/skttl/umbraco-matryoshka/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2137117257 | fix(Topology): :lady_beetle: Site_link not work for podman
Description
Briefly describe the changes made in this Pull Request.
Additional context (optional)
Add any additional context or information about the changes made in this PR.
How to test
Provide clear and concise instructions on how to test the changes made in this PR.
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Comparison is base (cc33451) 89.60% compared to head (2b3b257) 89.59%.
Additional details and impacted files
@@ Coverage Diff @@
## main #379 +/- ##
==========================================
- Coverage 89.60% 89.59% -0.02%
==========================================
Files 117 117
Lines 3455 3459 +4
Branches 774 775 +1
==========================================
+ Hits 3096 3099 +3
- Misses 329 330 +1
Partials 30 30
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
| gharchive/pull-request | 2024-02-15T17:55:09 | 2025-04-01T06:45:48.632509 | {
"authors": [
"bartoval",
"codecov-commenter"
],
"repo": "skupperproject/skupper-console",
"url": "https://github.com/skupperproject/skupper-console/pull/379",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1162806868 | Rename qpid-dispatch artifacts/binaries to skupper
Goal: avoid naming collisions with Apache Qpid Dispatch. It should be possible to install both skupper-router and qpid-dispatch-router on the same system without overlap.
Router Name changes
Binary
qdrouterd -> skrouterd
Tooling
qdstat -> skstat; qdmanage->skmanage
Python
python paths: qpid-dispatch -> skupper-router
Config file path (?)
/etc/qpid-dispatch/qdrouterd.conf -> /etc/skupper-router/skrouterd.conf
/etc/sasl2/qdrouterd.conf -> /etc/sasl2/skrouterd.conf
Other pathnames created during install:
/share/qpid-dispatch/
/include/qpid/dispatch/
/share/doc/qpid-dispatch/
/lib/python3.10/site-packages/qpid_dispatch[_site.py]/
/lib/qpid-dispatch/
Also, CMake project name in CMakeLists.txt: dispatch-router -> skupper-router
Thanks - got it.
| gharchive/issue | 2022-03-08T15:38:13 | 2025-04-01T06:45:48.637553 | {
"authors": [
"kgiusti"
],
"repo": "skupperproject/skupper-router",
"url": "https://github.com/skupperproject/skupper-router/issues/139",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1392102060 | [Question:] How to validate IBANs strict?
Hi @skwasjer,
thanks for implementing and maintaining this lib. I saw #24 and was wondering how to validate IBANs strict as you've mentioned in that ticket?
Inside Validate function there is always a NormalizeOrNull done:
https://github.com/skwasjer/IbanNet/blob/a1f7e1528ad033d1b8cd4b34bab79c4ca69259fb/src/IbanNet/IbanValidator.cs#L80
which converts to upper-casing:
https://github.com/skwasjer/IbanNet/blob/a1f7e1528ad033d1b8cd4b34bab79c4ca69259fb/src/IbanNet/Iban.cs#L258-L259
So I don't actually see any chance to validate strict (also validate if the casing is correct) unless to provide an additional Ruleto the ValidationOptions. (which feels like some sort of duplicating)
Maybe you have a better approach? Would there be a strict Flag in the Options possible?
Thanks you 😀
From what I remember, even though SWIFT has a distinction between upper vs lower case, there is currently no country that has IBAN's which require 'only' lowercase in certain positions (but perhaps I remember wrong).
In major versions prior to v4 (IIRC), this library would indeed check explicitly for lower and uppercase but that was 'negated' by the introduction of the normalize function as you mentioned in subsequent versions and I considered it fine because of that.
What is your use case for requiring support for this, for my understanding?
Because it could be somewhat impactful, notably the Iban type its formatting functions, equality checks, hash code, etc.
Btw, the strict vs loose validation debate back then is not entirely applicable anymore. In older versions you could choose either mode, where loose would skip certain checks (for performance). That was deprecated, as performance was improved and skipping the checks provided little gains.
And thinking some more, I think I may have misunderstood what you want to achieve, so do correct me. Do you just want to refuse validation when the input has lowercase but the IBAN pattern requires an upper case character?
@skwasjer thanks for your detailed explanation - indeed what I am trying to achieve is what you've asked in your last message:
Do you just want to fail validation when the input has lowercase but the IBAN pattern requires an upper case character
I need to do this because of a validation in a backend system which would fail if it is called with lower-case but upper-case is required.
In that case I would probably recommend (for now) using the IbanParser and Iban type in your domain/business logic. Internally it still uses the validator, but it allows you to make use of the strong typed API to ensure sanization/normalization and formatting:
var myInput = " nl91 ABna 0417 1643 00 ";
var ibanParser = new IbanParser(IbanRegistry.Default);
// Is valid?
if (ibanParser.TryParse(myInput, out Iban iban))
{
// Pass to backend system/DB/API, ensuring IBAN is formatted correctly, without whitespace, etc.
string formattedIbanStr = iban.ToString(IbanFormat.Electronic); // "NL91ABNA0417164300"
}
I do think it makes sense but would have to investigate further if we can do it by moving the sanitation logic 'only' into the parser and make the validator strict, without causing (breaking) side effects to the Iban type as mentioned previously.
@skwasjer thanks for your suggestion but that would force us to manually "translate" the values each time.
assume following infrastructure: we have a validation component which prevents invalid data to be placed into a queue. at the end of the queue sits a processor which forwards some of the payload to a thrid party component (which validates the iban (again)). actually we are not intercepting this process but with the conversion we have to (every time accessing and forwarding this attribute/property). second is, that with the validation sitting in front we could prevent invalid data coming into the system we have to deal with then. once it is in our system we have to handle all faults occuring. one more thing to consider - we actually don't won't to change the payload on our side - pretty much the same as we are responsible for any errors occuring in that step.
it would be great to decouple validation from parsing and have the possibility to validate more detailed. I think it makes absolutly sense once validation (weaker or stricter) is done and the iban is valid to transform in a more technical representation which gives the abilities it has nowadays with your library.
thanks!
Moving the sanitation to the parser is trivial, it is the breaking change/side effect that either needs to be documented well, or we have to figure out if it can be made opt-in as per your OP. To make matters more complicated, the side effects unfortunately would also affect for example the FluentValidator/DataAnnotations extensions library. I will give it some thought, and come back to you.
I've addressed your request in #95. Since it causes a behavior change, I will push it in next major soon-ish after I update docs and address some other things I want to improve.
Good morning Martijn,
that is great to hear - thank you so much for all your effort you put into this library 🙏🏼
v5.7.0
works perfectly - thanks Martijn 🙏🏼
| gharchive/issue | 2022-09-30T08:57:01 | 2025-04-01T06:45:48.649019 | {
"authors": [
"kimpenhaus",
"skwasjer"
],
"repo": "skwasjer/IbanNet",
"url": "https://github.com/skwasjer/IbanNet/issues/93",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
326215783 | Form renders success flash but email is never sent
Hi guys, I'm having an urgent issue - wasn't expecting it and my deadline is today :(
Whenever I send an email using this plugin, I get no errors whatsoever but emails are never sent. To be sure, I switched my credentials and am using gmail's - but the problem is the same.
Form
<form data-request="{{ __SELF__ }}::onFormSubmit" data-request-validate>
{{ form_token() }}
<div class="row">
<div id="{{ __SELF__ }}_forms_flash" class="col-md-12"></div>
</div>
<div class="row">
<div class="col-md-6">
<div class="form-group">
<input type="text" id="name" name="name" class="form-control" placeholder="Nome">
<div data-validate-for="name" class="has-error"></div>
</div>
</div>
<div class="col-md-6">
<div class="form-group">
<input type="text" id="email" name="email" class="form-control" placeholder="E-mail">
<div data-validate-for="email" class="has-error"></div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="form-group">
<textarea id="message" name="message" class="form-control" cols="30" rows="7" placeholder="Mensagem"></textarea>
<div data-validate-for="message" class="has-error"></div>
</div>
</div>
</div>
<div class="send">
<input id="submit" type="submit" value="Enviar" class="btn btn-primary">
</div>
</form>
.env file
MAIL_DRIVER=smtp (tried with sendmail too)
MAIL_HOST=smtp.gmail.com
MAIL_PORT=587
MAIL_USERNAME=myemail@gmail.com
MAIL_PASSWORD=mypassword
MAIL_ENCRYPTION=tls
OctoberCMS mail configuration
Whenever I send an email, the only thing I see in the XHR responses is this, which is weird - shouldn the POST response contain HTML?
Hope anyone can help me out. Thanks!
Sorry @sergiocastrovale but never used gmail smtp server...
Probably you need to use smtp as email method, but don't know the gmail config 😞
| gharchive/issue | 2018-05-24T17:19:05 | 2025-04-01T06:45:48.657424 | {
"authors": [
"sergiocastrovale",
"skydiver"
],
"repo": "skydiver/october-plugin-forms",
"url": "https://github.com/skydiver/october-plugin-forms/issues/115",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
230979551 | Magic Forms: Issue
How to fix this issue ?
"Array to string conversion" on line 179 of /var/www/html/octcms.dev/vendor/laravel/framework/src/Illuminate/Support/MessageBag.php
Hello @ashok2009it, can you post your form code?
Thanks.-
| gharchive/issue | 2017-05-24T09:40:26 | 2025-04-01T06:45:48.659715 | {
"authors": [
"ashok2009it",
"skydiver"
],
"repo": "skydiver/october-plugin-forms",
"url": "https://github.com/skydiver/october-plugin-forms/issues/59",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1984113483 | Archive this repository
This repository has become obsolete with Compose Compiler 1.5.4, which introduced stability configuration file.
@Nek-12 This library has supported KMP since version 1.0.2. Would you elaborate on your suggestion?
Yes, I'm sorry, I did not check the chamgelogs. Last time I checked it did not
Well, anyway, thanks for having an interest in this repository. I wish to archive this library one day once the Compose compiler supports stability marking across different independent modules.
But this issue was created exactly because compose added support for that...
| gharchive/issue | 2023-11-08T17:44:50 | 2025-04-01T06:45:48.661829 | {
"authors": [
"Nek-12",
"skydoves"
],
"repo": "skydoves/compose-stable-marker",
"url": "https://github.com/skydoves/compose-stable-marker/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
369784283 | [HELP] Is it possible to change the output file name for command "submission"??
When I run the command: leetcode submission 1, the generated output file name is two-sum.73790064.ac.cpp. I am wondering whether I can change the name such as two-sum.cpp (I just want to remove those series of numbers..)?
Thank you very much.
Don't think there is. At least for me I had to manually rename each file or you can simply write a script to do that.
Yep, recently I'm considering to move those meta into code file, instead of messing filename as it is.
Like to see that happen, e.g. python3 submission with py extension
Cool. Seems @skygragon is coming back to take care of this amazing but drowsy repo. Lol. You are away for a long time and we miss you. Really like leetcode-cli and really appreciate creating such a great tool. I hope this tool can be further polished.
| gharchive/issue | 2018-10-13T09:17:44 | 2025-04-01T06:45:48.664625 | {
"authors": [
"sfdye",
"skygragon",
"yanzhang0219"
],
"repo": "skygragon/leetcode-cli",
"url": "https://github.com/skygragon/leetcode-cli/issues/125",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
567392429 | [bug]Complied class has many useless variables
I had a test for this problem by using old version and latest version, the problem stays always.
I send the apk to my colleague, his result is good.
The follow is my running environment.
java version "1.8.0_201"
I hope you can solve it, i need Jadx for my work.
Thanks.
Please check the Jadx preferences. The posted code looks like you have enabled the "Fallback mode (simple dump) option.
Please check the Jadx preferences. The posted code looks like you have enabled the "Fallback mode (simple dump) option.
Yes, thats's the problem. You save me lots of time. Thank you very much.
| gharchive/issue | 2020-02-19T08:17:29 | 2025-04-01T06:45:48.669512 | {
"authors": [
"GvcZhang",
"jpstotz"
],
"repo": "skylot/jadx",
"url": "https://github.com/skylot/jadx/issues/859",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1355617728 | View on all entities except some with specific component
Hi, It there a way to grap all entities that do not have some component like:
registry.view(exclude) ?
At the moment, you can only iterate all entities (registry::each) and if them explicitly.
However, what you're asking for is on the roadmap. This is the final interface fyi:
auto view = registry.view<void>(entt::exclude<T, U>);
Great to know. From a performance perspective, is it more efficient to create a BaseComponent to all entity and query it with my exclude, or the each + if on any_of ?
If you create a base component and exlude it, it doesn't mean that all derived classes are also excluded as a consequence. I guess this answers your question.
I meaned a struct BaseComponent that ensure that every entities have one (each time i create an entiti, I add this BaseComponent to it)
then, I can:
auto view = Registry.view<BaseComponent>(exclude<ComponentToExclue>);
I would use this BaseComponent as a workaround to get a view to all my entities
Oh, sorry, didn't get it. Then I would go with the BaseComponent since it will be easier to update later on (just replace BaseComponent with void and you're done).
Great. BTW, that lead to a feature request : is there any way to get a signal on a new entity creation (and destruction) as we have with components instance
Available upstream on the wip branch. The way it works is:
for(auto entt: registry.view<entt::entity>(entt::exclude<T>)) { ... }
Signals on entity construction/deletion are also available as in:
registry.on_construct<entt::entity>().connect<&my_func>();
Available on master. :+1:
I've just noticed that the size() of this kind of view is not available:
auto myview = registry.viewentt::entity();
myview.size() returns the total number of entities
auto myview = registry.viewentt::entity(entt::exclude);
myview.size() -> ERROR : size() not a member of ...
Is there any workaround ?
size_hint @christopheseyler it tells you the maximum number of elements returned by the view.
It's not guaranteed to return exactly N elements but it never returns more than them.
| gharchive/issue | 2022-08-30T11:31:15 | 2025-04-01T06:45:48.680544 | {
"authors": [
"christopheseyler",
"skypjack"
],
"repo": "skypjack/entt",
"url": "https://github.com/skypjack/entt/issues/929",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1412674049 | Sky camp modifications to support CLI guardrails
Skyplane VPC should allow connections to all ports and IPs
Disable reuse_gateways for transfers
Configure deprovision to scope it to a user’s UUID
Disable GCP and Azure support in Skyplane init
@ShishirPatil Please test this PR, it adds some guardrails for participants on the skycamp-tutorial branch.
Disregard GCP and Azure unit tests as I disable both for Sky camp
| gharchive/pull-request | 2022-10-18T06:57:54 | 2025-04-01T06:45:48.682630 | {
"authors": [
"parasj"
],
"repo": "skyplane-project/skyplane",
"url": "https://github.com/skyplane-project/skyplane/pull/618",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2545969501 | Pasting an URL that has been copied from the browser doesn't trigger any clipboard matcher
If you add matchers with the clipboard module, they are not triggered when pasting an URL that has been copied from the URL bar of the navigator (Google Chrome on Mac).
If you type a valid URL in the editor, then select it, copy and paste, the matcher is triggered.
Steps for Reproduction
Visit quilljs.com
Copy the URL from any opened tab of your browser (can be the actual tab the you are at the moment).
Paste in editor.
Expected behavior:
The following text should be inserted in the editor: 'I just pasted this: ' + (the actual URL copied).
Plus, in the console, there should be a log saying: 'OK. This is a text node' or 'OK. This is an element node'.
Actual behavior:
None of the matchers are called. But the URL appears in the editor (which is not the wanted behavior, we want to catch the URL and format it).
Platforms:
Mac OS (14.3.1), Chrome for MAC (128.0.6613.138).
Version:
2.0.2
Run Quill.version to find out
Duplicate of #4406
Workaround
quill.container.addEventListener('paste', ev => {
const text = ev.clipboardData.getData('text/plain');
ev.preventDefault();
const html = `<p>${text}</p>`;
const clipboard = quill.getModule('clipboard');
clipboard.dangerouslyPasteHTML(quill.getSelection().index, html);
}, true);
Another workaround, slightly better (doesn't touch handler if it has html already)
quill.root.addEventListener(
'paste',
(ev) => {
const text = ev.clipboardData.getData('text/plain')
const html = ev.clipboardData.getData('text/html')
if (!html && text) {
ev.preventDefault()
// capture event with fixed content by quill. Sadly clipboardData.setData doesn't seem to work.
const data = new DataTransfer()
data.setData('text/plain', text)
data.setData('text/html', `<p>${text}</p>`)
const event = new ClipboardEvent('paste', { clipboardData: data })
quill.clipboard.onCapturePaste(event)
}
},
true
)
| gharchive/issue | 2024-09-24T17:30:18 | 2025-04-01T06:45:48.689863 | {
"authors": [
"Valian",
"compmaster",
"jpbecotte"
],
"repo": "slab/quill",
"url": "https://github.com/slab/quill/issues/4421",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1650379638 | Trying to post message in channel as well as DM's using slash commands
So, I was trying to recreate an app called giphy but serverless, while implementing the code, I am facing a issue on how a slash command can be used in channels and direct message simultaneously. And another thing which giphy does is, it uses username and avatar pic of the user which uses the slash command irrespective of using it in channels or dms. I tried I was able to replicate but when I used xoxb token I get the reply from BOT (avatar and username of bot) but when I try using xoxp token I get desired result but if different user uses the slash command it shows my avatar and username only.
Reproducible in:
The Slack SDK version
Slack Bolt: 3.12.2
Node.js runtime version
nodejs16.x
OS info
macOS: 13.2.1
Steps to reproduce:
I am trying to send the gif on a click of a button.
below is the code that will help to reproduce it.
Expected result:
When using the slash command in channel or direct messages,
then the user who triggered the slash command able to send the gif using his avatar and username
Actual result:
As of now the creator of the workspace avatar and username is being used irrespective of who triggered the slash command.
Requirements:
AWS SAM, slash command
For general questions/issues about Slack API platform or its server-side, could you submit questions at https://my.slack.com/help/requests/new instead. :bow:
Please read the Contributing guidelines and Code of Conduct before creating this issue or pull request. By submitting, you are agreeing to those rules.
What scopes does your app have? It is possible that the Giphy app is a legacy app and therefore has more permissions than an app we can build today
Hi @Shadow-Flash , thanks for asking the question.
As @jeffbaldwinjr mentioned, with the latest permission model, you need to have an additional scope called chat:write.customize to pass username and icon properties to chat.postMessage API calls.
So, as long as your app can perform chat.postMessage API, customizing the appearance is still possible. Contrarily, when your app's slash command is invoked in a person's DM without anyone else or in a private channel your app's bot user is not a member of, your app cannot perform a chat.postMessage API call there. With the past permission model, there used be a way to post a message, but it's no longer supported. Thus, in this scenario, there is no way to customize the apperance. In other words, when your app needs to reply using ack() or respond(), username and icon_* do not work.
This may not be a great answer for you but I hope this helps.
@jeffbaldwinjr I have the necessary scopes defined in Scopes Section for Bot token specifically (I will mention them here for future reference: chat:write, chat:write.customize). And I didn't think through that giphy app can be a legacy app 😓 .
@seratch Yeah so, trying to duplicate the features of giphy app won't be my goal anymore 😄.
Thanks both of you (@jeffbaldwinjr and @seratch )for clearing my doubt 💯 . I will close this issue now.
| gharchive/issue | 2023-04-01T09:11:56 | 2025-04-01T06:45:48.700082 | {
"authors": [
"Shadow-Flash",
"jeffbaldwinjr",
"seratch"
],
"repo": "slackapi/bolt-js",
"url": "https://github.com/slackapi/bolt-js/issues/1793",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
207690868 | adds example application
greet and react is an example app that shows off a fully functional application that uses this package.
Requirements (place an x in each [ ])
[x] I've read and understood the Contributing guidelines and have done my best effort to follow them.
[x] I've read and agree to the Code of Conduct.
[x] I've been mindful about doing atomic commits, adding documentation to my changes, not refactoring too much.
[x] I've a descriptive title and added any useful information for the reviewer. Where appropriate, I've attached a screenshot and/or screencast (gif preferably).
[x] I've written tests to cover the new code and functionality included in this PR.
[x] I've read, agree to, and signed the Contributor License Agreement (CLA).
i honestly don't know what it takes to get codecov reporting once it gets into this weird state.
oh cool it works!
| gharchive/pull-request | 2017-02-15T02:13:04 | 2025-04-01T06:45:48.707573 | {
"authors": [
"aoberoi"
],
"repo": "slackapi/node-slack-events-api",
"url": "https://github.com/slackapi/node-slack-events-api/pull/11",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
782602485 | Does not connect to MQTT Broker..
HASS Workstation Service will not connect to broker.
Which port number is being used fir MQTT?
Can i enter a port number on "IP or Hostname" field?
I can connect to my broker with mqtt://XX.XX.XX.XX:8883 + [user] & [password]
Hi @ChrisPiel70!
The app does not support any ports except 1883 yet and it won't work when you add the port number in the "IP or Hostname" field.
Hi @ChrisPiel70!
The app does not support any ports except 1883 yet and it won't work when you add the port number in the "IP or Hostname" field.
This shouldn't be too much work so I'll get on it.
This shouldn't be too much work so I'll get on it.
This has been implemented and will be in the next release! This will be later today or tomorrow.
This has been implemented and will be in the next release! This will be later today or tomorrow.
| gharchive/issue | 2021-01-09T12:54:03 | 2025-04-01T06:45:48.744927 | {
"authors": [
"ChrisPiel70",
"sleevezipper"
],
"repo": "sleevezipper/hass-workstation-service",
"url": "https://github.com/sleevezipper/hass-workstation-service/issues/11",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1040430340 | Printer job Sensor
A sensor that indicates whether a printer job is present would be desirable. Possibly also the name of the job, but a true/false to the job is already good enough.
Thus one can make automation, which switches the printer on, if a printer job is present.
this might help https://www.home-assistant.io/integrations/cups/
| gharchive/issue | 2021-10-31T10:19:39 | 2025-04-01T06:45:48.746331 | {
"authors": [
"Diddlik",
"SecretiveShell"
],
"repo": "sleevezipper/hass-workstation-service",
"url": "https://github.com/sleevezipper/hass-workstation-service/issues/119",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.